sh-4.4# journalctl -u kubelet -u crio --nopagero[1@-[1@  -- Logs begin at Thu 2023-01-05 07:14:41 UTC, end at Thu 2023-01-05 17:46:39 UTC. -- Jan 05 15:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 15:27:34.754386481Z" level=info msg="Starting CRI-O, version: 1.25.1-5.rhaos4.12.git6005903.el8, git: unknown(clean)" Jan 05 15:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 15:27:34.754601316Z" level=info msg="Node configuration value for hugetlb cgroup is true" Jan 05 15:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 15:27:34.754616504Z" level=info msg="Node configuration value for pid cgroup is true" Jan 05 15:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 15:27:34.754683647Z" level=info msg="Node configuration value for memoryswap cgroup is true" Jan 05 15:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 15:27:34.754693493Z" level=info msg="Node configuration value for cgroup v2 is false" Jan 05 15:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 15:27:34.767799390Z" level=info msg="Node configuration value for systemd CollectMode is true" Jan 05 15:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 15:27:34.777622472Z" level=info msg="Node configuration value for systemd AllowedCPUs is true" Jan 05 15:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 15:27:34.779996728Z" level=info msg="Using default capabilities: CAP_CHOWN, CAP_DAC_OVERRIDE, CAP_FSETID, CAP_FOWNER, CAP_SETGID, CAP_SETUID, CAP_SETPCAP, CAP_NET_BIND_SERVICE, CAP_KILL" Jan 05 15:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Starting Container Runtime Interface for OCI (CRI-O)... Jan 05 15:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 15:27:34.873779381Z" level=info msg="Checkpoint/restore support disabled" Jan 05 15:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 15:27:34.873910209Z" level=info msg="Using seccomp default profile when unspecified: true" Jan 05 15:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 15:27:34.874000964Z" level=info msg="Using the internal default seccomp profile" Jan 05 15:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 15:27:34.874063571Z" level=info msg="AppArmor is disabled by the system or at CRI-O build-time" Jan 05 15:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 15:27:34.874111590Z" level=info msg="No blockio config file specified, blockio not configured" Jan 05 15:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 15:27:34.874188078Z" level=info msg="RDT not available in the host system" Jan 05 15:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 15:27:34.879848375Z" level=info msg="Conmon does support the --sync option" Jan 05 15:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 15:27:34.879867462Z" level=info msg="Conmon does support the --log-global-size-max option" Jan 05 15:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 15:27:34.883547329Z" level=info msg="Conmon does support the --sync option" Jan 05 15:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 15:27:34.883565300Z" level=info msg="Conmon does support the --log-global-size-max option" Jan 05 15:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 15:27:34.884012510Z" level=info msg="Updated default CNI network name to " Jan 05 15:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 15:27:34.925113956Z" level=warning msg="Error encountered when checking whether cri-o should wipe images: open /var/lib/crio/version: no such file or directory" Jan 05 15:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 15:27:34.927706679Z" level=info msg="Serving metrics on :9537 via HTTP" Jan 05 15:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Started Container Runtime Interface for OCI (CRI-O). Jan 05 15:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Starting Kubernetes Kubelet... Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: Flag --container-runtime has been deprecated, will be removed in 1.27 as the only valid value is 'remote' Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: Flag --cloud-provider has been deprecated, will be removed in 1.25 or later, in favor of removing cloud provider code from Kubelet. Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: Flag --provider-id has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.27. Image garbage collector will get sandbox image information from CRI. Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.318039 1425 server.go:200] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: Flag --container-runtime has been deprecated, will be removed in 1.27 as the only valid value is 'remote' Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: Flag --cloud-provider has been deprecated, will be removed in 1.25 or later, in favor of removing cloud provider code from Kubelet. Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: Flag --provider-id has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.27. Image garbage collector will get sandbox image information from CRI. Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.319835 1425 flags.go:64] FLAG: --add-dir-header="false" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.319857 1425 flags.go:64] FLAG: --address="0.0.0.0" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.319866 1425 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.319880 1425 flags.go:64] FLAG: --alsologtostderr="false" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.319887 1425 flags.go:64] FLAG: --anonymous-auth="true" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.319895 1425 flags.go:64] FLAG: --application-metrics-count-limit="100" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.319900 1425 flags.go:64] FLAG: --authentication-token-webhook="false" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.319905 1425 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.319912 1425 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.319918 1425 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.319923 1425 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.319929 1425 flags.go:64] FLAG: --azure-container-registry-config="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.319934 1425 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.319940 1425 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.319946 1425 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.319951 1425 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.319957 1425 flags.go:64] FLAG: --cgroup-root="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.319963 1425 flags.go:64] FLAG: --cgroups-per-qos="true" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.319968 1425 flags.go:64] FLAG: --client-ca-file="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.319973 1425 flags.go:64] FLAG: --cloud-config="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.319978 1425 flags.go:64] FLAG: --cloud-provider="external" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.319986 1425 flags.go:64] FLAG: --cluster-dns="[]" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.319996 1425 flags.go:64] FLAG: --cluster-domain="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320002 1425 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320007 1425 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320013 1425 flags.go:64] FLAG: --container-log-max-files="5" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320020 1425 flags.go:64] FLAG: --container-log-max-size="10Mi" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320025 1425 flags.go:64] FLAG: --container-runtime="remote" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320030 1425 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320036 1425 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320042 1425 flags.go:64] FLAG: --containerd-namespace="k8s.io" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320047 1425 flags.go:64] FLAG: --contention-profiling="false" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320052 1425 flags.go:64] FLAG: --cpu-cfs-quota="true" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320057 1425 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320063 1425 flags.go:64] FLAG: --cpu-manager-policy="none" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320068 1425 flags.go:64] FLAG: --cpu-manager-policy-options="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320075 1425 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320080 1425 flags.go:64] FLAG: --enable-controller-attach-detach="true" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320088 1425 flags.go:64] FLAG: --enable-debugging-handlers="true" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320094 1425 flags.go:64] FLAG: --enable-load-reader="false" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320104 1425 flags.go:64] FLAG: --enable-server="true" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320110 1425 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320122 1425 flags.go:64] FLAG: --event-burst="10" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320127 1425 flags.go:64] FLAG: --event-qps="5" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320133 1425 flags.go:64] FLAG: --event-storage-age-limit="default=0" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320161 1425 flags.go:64] FLAG: --event-storage-event-limit="default=0" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320166 1425 flags.go:64] FLAG: --eviction-hard="imagefs.available<15%,memory.available<100Mi,nodefs.available<10%,nodefs.inodesFree<5%" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320182 1425 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320187 1425 flags.go:64] FLAG: --eviction-minimum-reclaim="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320192 1425 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320198 1425 flags.go:64] FLAG: --eviction-soft="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320203 1425 flags.go:64] FLAG: --eviction-soft-grace-period="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320208 1425 flags.go:64] FLAG: --exit-on-lock-contention="false" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320214 1425 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320221 1425 flags.go:64] FLAG: --experimental-mounter-path="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320227 1425 flags.go:64] FLAG: --fail-swap-on="true" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320232 1425 flags.go:64] FLAG: --feature-gates="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320239 1425 flags.go:64] FLAG: --file-check-frequency="20s" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320245 1425 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320253 1425 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320260 1425 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320265 1425 flags.go:64] FLAG: --healthz-port="10248" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320271 1425 flags.go:64] FLAG: --help="false" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320276 1425 flags.go:64] FLAG: --hostname-override="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320281 1425 flags.go:64] FLAG: --housekeeping-interval="10s" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320286 1425 flags.go:64] FLAG: --http-check-frequency="20s" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320292 1425 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320297 1425 flags.go:64] FLAG: --image-credential-provider-config="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320302 1425 flags.go:64] FLAG: --image-gc-high-threshold="85" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320307 1425 flags.go:64] FLAG: --image-gc-low-threshold="80" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320314 1425 flags.go:64] FLAG: --image-service-endpoint="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320319 1425 flags.go:64] FLAG: --iptables-drop-bit="15" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320324 1425 flags.go:64] FLAG: --iptables-masquerade-bit="14" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320330 1425 flags.go:64] FLAG: --keep-terminated-pod-volumes="false" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320335 1425 flags.go:64] FLAG: --kernel-memcg-notification="false" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320340 1425 flags.go:64] FLAG: --kube-api-burst="10" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320345 1425 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320351 1425 flags.go:64] FLAG: --kube-api-qps="5" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320361 1425 flags.go:64] FLAG: --kube-reserved="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320366 1425 flags.go:64] FLAG: --kube-reserved-cgroup="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320371 1425 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320377 1425 flags.go:64] FLAG: --kubelet-cgroups="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320382 1425 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320388 1425 flags.go:64] FLAG: --lock-file="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320393 1425 flags.go:64] FLAG: --log-backtrace-at=":0" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320399 1425 flags.go:64] FLAG: --log-cadvisor-usage="false" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320406 1425 flags.go:64] FLAG: --log-dir="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320411 1425 flags.go:64] FLAG: --log-file="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320416 1425 flags.go:64] FLAG: --log-file-max-size="1800" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320422 1425 flags.go:64] FLAG: --log-flush-frequency="5s" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320427 1425 flags.go:64] FLAG: --log-json-info-buffer-size="0" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320436 1425 flags.go:64] FLAG: --log-json-split-stream="false" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320441 1425 flags.go:64] FLAG: --logging-format="text" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320447 1425 flags.go:64] FLAG: --logtostderr="true" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320467 1425 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320475 1425 flags.go:64] FLAG: --make-iptables-util-chains="true" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320480 1425 flags.go:64] FLAG: --manifest-url="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320485 1425 flags.go:64] FLAG: --manifest-url-header="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320496 1425 flags.go:64] FLAG: --master-service-namespace="default" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320502 1425 flags.go:64] FLAG: --max-housekeeping-interval="15s" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320507 1425 flags.go:64] FLAG: --max-open-files="1000000" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320513 1425 flags.go:64] FLAG: --max-pods="110" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320521 1425 flags.go:64] FLAG: --maximum-dead-containers="-1" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320526 1425 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320531 1425 flags.go:64] FLAG: --memory-manager-policy="None" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320536 1425 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320542 1425 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320547 1425 flags.go:64] FLAG: --node-ip="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320552 1425 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/worker=,node.openshift.io/os_id=rhcos" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320562 1425 flags.go:64] FLAG: --node-status-max-images="50" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320567 1425 flags.go:64] FLAG: --node-status-update-frequency="10s" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320572 1425 flags.go:64] FLAG: --one-output="false" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320578 1425 flags.go:64] FLAG: --oom-score-adj="-999" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320583 1425 flags.go:64] FLAG: --pod-cidr="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320589 1425 flags.go:64] FLAG: --pod-infra-container-image="registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320597 1425 flags.go:64] FLAG: --pod-manifest-path="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320602 1425 flags.go:64] FLAG: --pod-max-pids="-1" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320612 1425 flags.go:64] FLAG: --pods-per-core="0" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320619 1425 flags.go:64] FLAG: --port="10250" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320625 1425 flags.go:64] FLAG: --protect-kernel-defaults="false" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320630 1425 flags.go:64] FLAG: --provider-id="alicloud://us-east-1.i-0xif5dui7khr4rm1wp68" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320636 1425 flags.go:64] FLAG: --qos-reserved="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320642 1425 flags.go:64] FLAG: --read-only-port="10255" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320647 1425 flags.go:64] FLAG: --register-node="true" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320652 1425 flags.go:64] FLAG: --register-schedulable="true" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320661 1425 flags.go:64] FLAG: --register-with-taints="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320667 1425 flags.go:64] FLAG: --registry-burst="10" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320672 1425 flags.go:64] FLAG: --registry-qps="5" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320677 1425 flags.go:64] FLAG: --reserved-cpus="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320682 1425 flags.go:64] FLAG: --reserved-memory="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320689 1425 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320694 1425 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320699 1425 flags.go:64] FLAG: --rotate-certificates="false" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320705 1425 flags.go:64] FLAG: --rotate-server-certificates="false" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320711 1425 flags.go:64] FLAG: --runonce="false" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320717 1425 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320723 1425 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320733 1425 flags.go:64] FLAG: --seccomp-default="false" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320743 1425 flags.go:64] FLAG: --serialize-image-pulls="true" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320749 1425 flags.go:64] FLAG: --skip-headers="false" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320754 1425 flags.go:64] FLAG: --skip-log-headers="false" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320759 1425 flags.go:64] FLAG: --stderrthreshold="2" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320764 1425 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320770 1425 flags.go:64] FLAG: --storage-driver-db="cadvisor" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320775 1425 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320781 1425 flags.go:64] FLAG: --storage-driver-password="root" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320786 1425 flags.go:64] FLAG: --storage-driver-secure="false" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320791 1425 flags.go:64] FLAG: --storage-driver-table="stats" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320796 1425 flags.go:64] FLAG: --storage-driver-user="root" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320801 1425 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320807 1425 flags.go:64] FLAG: --sync-frequency="1m0s" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320816 1425 flags.go:64] FLAG: --system-cgroups="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320821 1425 flags.go:64] FLAG: --system-reserved="cpu=500m,ephemeral-storage=1Gi,memory=1Gi" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320830 1425 flags.go:64] FLAG: --system-reserved-cgroup="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320835 1425 flags.go:64] FLAG: --tls-cert-file="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320840 1425 flags.go:64] FLAG: --tls-cipher-suites="[]" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320855 1425 flags.go:64] FLAG: --tls-min-version="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320861 1425 flags.go:64] FLAG: --tls-private-key-file="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320866 1425 flags.go:64] FLAG: --topology-manager-policy="none" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320871 1425 flags.go:64] FLAG: --topology-manager-scope="container" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320876 1425 flags.go:64] FLAG: --v="2" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320883 1425 flags.go:64] FLAG: --version="false" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320890 1425 flags.go:64] FLAG: --vmodule="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320896 1425 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.320902 1425 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.321024 1425 feature_gate.go:246] feature gates: &{map[APIPriorityAndFairness:true CSIMigrationAzureFile:false CSIMigrationvSphere:false DownwardAPIHugePages:true RotateKubeletServerCertificate:true]} Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.325439 1425 server.go:413] "Kubelet version" kubeletVersion="v1.25.4+77bec7a" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.325476 1425 server.go:415] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.325535 1425 feature_gate.go:246] feature gates: &{map[APIPriorityAndFairness:true CSIMigrationAzureFile:false CSIMigrationvSphere:false DownwardAPIHugePages:true RotateKubeletServerCertificate:true]} Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.325625 1425 feature_gate.go:246] feature gates: &{map[APIPriorityAndFairness:true CSIMigrationAzureFile:false CSIMigrationvSphere:false DownwardAPIHugePages:true RotateKubeletServerCertificate:true]} Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.325735 1425 server.go:825] "Client rotation is on, will bootstrap in background" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.330892 1425 bootstrap.go:100] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.333419 1425 server.go:882] "Starting client certificate rotation" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.333482 1425 certificate_manager.go:270] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.333659 1425 certificate_manager.go:270] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.340995 1425 dynamic_cafile_content.go:119] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.342759 1425 manager.go:163] cAdvisor running in container: "/system.slice/kubelet.service" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.343046 1425 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.347391 1425 fs.go:133] Filesystem UUIDs: map[249B-6B8C:/dev/vda2 7e283ec8-4fd4-48d9-bedf-150e5fbbb374:/dev/vda3 fda260a9-581c-4391-b124-5cfcba02ef65:/dev/vda4] Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.347407 1425 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /sys/fs/cgroup:{mountpoint:/sys/fs/cgroup major:0 minor:25 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:43 fsType:tmpfs blockSize:0}] Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.347441 1425 nvidia.go:54] NVIDIA GPU metrics disabled Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.349278 1425 manager.go:212] Machine: {Timestamp:2023-01-05 15:27:35.349090195 +0000 UTC m=+0.318509894 CPUVendorID:GenuineIntel NumCores:2 NumPhysicalCores:1 NumSockets:1 CpuFrequency:2500000 MemoryCapacity:8192086016 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:2bf656612c3249b68a06a3657cb09103 SystemUUID:2bf65661-2c32-49b6-8a06-a3657cb09103 BootID:d563ac77-827a-484b-9fd1-80a5dd5aaf96 Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:4096040960 Type:vfs Inodes:1000010 HasInodes:true} {Device:/sys/fs/cgroup DeviceMajor:0 DeviceMinor:25 Capacity:4096040960 Type:vfs Inodes:1000010 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:128300593152 Type:vfs Inodes:62651840 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:43 Capacity:4096040960 Type:vfs Inodes:1000010 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:4096040960 Type:vfs Inodes:1000010 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:128849018880 Scheduler:mq-deadline}] NetworkDevices:[{Name:ens5 MacAddress:00:16:3e:00:58:a6 Speed:-1 Mtu:1500}] Topology:[{Id:0 Memory:8192086016 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0 1] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0}] Caches:[{Id:0 Size:37486592 Type:Unified Level:3}]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.349420 1425 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.349630 1425 manager.go:228] Version: {KernelVersion:4.18.0-372.39.1.el8_6.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 412.86.202212170457-0 (Ootpa) DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.350060 1425 container_manager_linux.go:262] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.350173 1425 container_manager_linux.go:267] "Creating Container Manager object based on Node Config" nodeConfig={RuntimeCgroupsName:/system.slice/crio.service SystemCgroupsName:/system.slice KubeletCgroupsName: KubeletOOMScoreAdj:-999 ContainerRuntime: CgroupsPerQOS:true CgroupRoot:/ CgroupDriver:systemd KubeletRootDir:/var/lib/kubelet ProtectKernelDefaults:false NodeAllocatableConfig:{KubeReservedCgroupName: SystemReservedCgroupName: ReservedSystemCPUs: EnforceNodeAllocatable:map[pods:{}] KubeReserved:map[] SystemReserved:map[cpu:{i:{value:500 scale:-3} d:{Dec:} s:500m Format:DecimalSI} ephemeral-storage:{i:{value:1073741824 scale:0} d:{Dec:} s:1Gi Format:BinarySI} memory:{i:{value:1073741824 scale:0} d:{Dec:} s:1Gi Format:BinarySI}] HardEvictionThresholds:[{Signal:nodefs.inodesFree Operator:LessThan Value:{Quantity: Percentage:0.05} GracePeriod:0s MinReclaim:} {Signal:imagefs.available Operator:LessThan Value:{Quantity: Percentage:0.15} GracePeriod:0s MinReclaim:} {Signal:memory.available Operator:LessThan Value:{Quantity:100Mi Percentage:0} GracePeriod:0s MinReclaim:} {Signal:nodefs.available Operator:LessThan Value:{Quantity: Percentage:0.1} GracePeriod:0s MinReclaim:}]} QOSReserved:map[] ExperimentalCPUManagerPolicy:none ExperimentalCPUManagerPolicyOptions:map[] ExperimentalTopologyManagerScope:container ExperimentalCPUManagerReconcilePeriod:10s ExperimentalMemoryManagerPolicy:None ExperimentalMemoryManagerReservedMemory:[] ExperimentalPodPidsLimit:4096 EnforceCPULimits:true CPUCFSQuotaPeriod:100ms ExperimentalTopologyManagerPolicy:none} Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.350199 1425 topology_manager.go:134] "Creating topology manager with policy per scope" topologyPolicyName="none" topologyScopeName="container" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.350214 1425 container_manager_linux.go:302] "Creating device plugin manager" devicePluginEnabled=true Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.350773 1425 manager.go:127] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.350792 1425 server.go:64] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.351088 1425 state_mem.go:36] "Initialized new in-memory state store" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.351169 1425 util_unix.go:104] "Using this endpoint is deprecated, please consider using full URL format" endpoint="/var/run/crio/crio.sock" URL="unix:///var/run/crio/crio.sock" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.370231 1425 remote_runtime.go:139] "Using CRI v1 runtime API" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.370370 1425 util_unix.go:104] "Using this endpoint is deprecated, please consider using full URL format" endpoint="/var/run/crio/crio.sock" URL="unix:///var/run/crio/crio.sock" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.374653 1425 remote_image.go:95] "Using CRI v1 image API" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.374682 1425 server.go:1136] "Using root directory" path="/var/lib/kubelet" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.376443 1425 kubelet.go:393] "Attempting to sync node with API server" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.376528 1425 kubelet.go:282] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.376560 1425 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.376575 1425 kubelet.go:293] "Adding apiserver pod source" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.376597 1425 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.378895 1425 kuberuntime_manager.go:240] "Container runtime initialized" containerRuntime="cri-o" version="1.25.1-5.rhaos4.12.git6005903.el8" apiVersion="v1" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.382638 1425 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.382657 1425 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/rbd" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.382668 1425 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/aws-ebs" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.382681 1425 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/gce-pd" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.382693 1425 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/cinder" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.382704 1425 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/azure-disk" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.382715 1425 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/azure-file" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.382727 1425 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/vsphere-volume" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.383114 1425 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.383132 1425 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.383166 1425 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.383177 1425 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.383188 1425 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/secret" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.383205 1425 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.383216 1425 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/glusterfs" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.383228 1425 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/cephfs" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.383241 1425 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.383256 1425 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/fc" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.383269 1425 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Started Kubernetes Kubelet. Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.383281 1425 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/projected" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.383293 1425 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.383333 1425 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/csi" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.383520 1425 server.go:1175] "Started kubelet" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:35.385443 1425 kubelet.go:1333] "Image garbage collection failed once. Stats initialization may not have completed yet" err="failed to get imageFs info: unable to find data in memory cache" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 15:27:35.387664 1425 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Node: nodes "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:35.387780 1425 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Node: failed to list *v1.Node: nodes "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 15:27:35.387893 1425 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:35.387949 1425 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.388026 1425 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.390214 1425 server.go:155] "Starting to listen" address="0.0.0.0" port=10250 Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.390933 1425 csr.go:261] certificate signing request csr-skclg is approved, waiting to be issued Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.393130 1425 server.go:438] "Adding debug handlers to kubelet server" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.399115 1425 csr.go:257] certificate signing request csr-skclg is issued Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:35.401226 1425 event.go:267] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.17377334605a414d", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 383449933, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 383449933, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'events is forbidden: User "system:anonymous" cannot create resource "events" in API group "" in the namespace "default"' (will not retry!) Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.401804 1425 certificate_manager.go:270] kubernetes.io/kubelet-serving: Certificate rotation is enabled Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 15:27:35.402575474Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=625f0ff5-53c7-40dc-9bcf-03ee9f80f97e name=/runtime.v1.ImageService/ImageStatus Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 15:27:35.405059873Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12 not found" id=625f0ff5-53c7-40dc-9bcf-03ee9f80f97e name=/runtime.v1.ImageService/ImageStatus Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.405436 1425 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.405895 1425 volume_manager.go:291] "The desired_state_of_world populator starts" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.405997 1425 volume_manager.go:293] "Starting Kubelet Volume Manager" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.406387 1425 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:35.409432 1425 kubelet.go:2396] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:35.409684 1425 controller.go:144] failed to ensure lease exists, will retry in 200ms, error: leases.coordination.k8s.io "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" is forbidden: User "system:anonymous" cannot get resource "leases" in API group "coordination.k8s.io" in the namespace "kube-node-lease" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 15:27:35.409914 1425 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:35.409942 1425 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.411898 1425 factory.go:153] Registering CRI-O factory Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.411914 1425 factory.go:55] Registering systemd factory Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.413990 1425 factory.go:103] Registering Raw factory Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.414613 1425 manager.go:1201] Started watching for new ooms in manager Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.418082 1425 manager.go:302] Starting recovery of all containers Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.479102 1425 manager.go:307] Recovery completed Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.513248 1425 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:35.517530 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.543980 1425 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.544022 1425 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.544036 1425 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.544074 1425 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:35.547963 1425 kubelet_node_status.go:94] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:35.548890 1425 event.go:267] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ec0bb0", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientMemory", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientMemory", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 543999408, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 543999408, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'events is forbidden: User "system:anonymous" cannot create resource "events" in API group "" in the namespace "default"' (will not retry!) Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:35.549946 1425 event.go:267] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ec7a97", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasNoDiskPressure", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasNoDiskPressure", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 544027799, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 544027799, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'events is forbidden: User "system:anonymous" cannot create resource "events" in API group "" in the namespace "default"' (will not retry!) Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:35.550931 1425 event.go:267] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ecab0e", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientPID", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientPID", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 544040206, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 544040206, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'events is forbidden: User "system:anonymous" cannot create resource "events" in API group "" in the namespace "default"' (will not retry!) Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.595961 1425 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.597531 1425 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.597582 1425 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.597596 1425 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.598921 1425 cpu_manager.go:214] "Starting CPU manager" policy="none" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.598936 1425 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.598953 1425 state_mem.go:36] "Initialized new in-memory state store" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:35.599445 1425 event.go:267] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ec0bb0", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientMemory", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientMemory", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 543999408, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 597561707, time.Local), Count:2, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'events "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ec0bb0" is forbidden: User "system:anonymous" cannot patch resource "events" in API group "" in the namespace "default"' (will not retry!) Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:35.600774 1425 event.go:267] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ec7a97", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasNoDiskPressure", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasNoDiskPressure", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 544027799, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 597587690, time.Local), Count:2, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'events "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ec7a97" is forbidden: User "system:anonymous" cannot patch resource "events" in API group "" in the namespace "default"' (will not retry!) Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.601101 1425 policy_none.go:49] "None policy: Start" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.601637 1425 memory_manager.go:168] "Starting memorymanager" policy="None" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.601664 1425 state_mem.go:35] "Initializing new in-memory state store" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:35.602263 1425 event.go:267] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ecab0e", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientPID", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientPID", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 544040206, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 597601292, time.Local), Count:2, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'events "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ecab0e" is forbidden: User "system:anonymous" cannot patch resource "events" in API group "" in the namespace "default"' (will not retry!) Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.602977 1425 container_manager_linux.go:427] "Updating kernel flag" flag="vm/overcommit_memory" expectedValue=1 actualValue=0 Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.603065 1425 container_manager_linux.go:427] "Updating kernel flag" flag="kernel/panic" expectedValue=10 actualValue=0 Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:35.612695 1425 controller.go:144] failed to ensure lease exists, will retry in 400ms, error: leases.coordination.k8s.io "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" is forbidden: User "system:anonymous" cannot get resource "leases" in API group "coordination.k8s.io" in the namespace "kube-node-lease" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:35.617804 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.633758 1425 kubelet_network_linux.go:63] "Initialized iptables rules." protocol=IPv4 Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.638342 1425 manager.go:273] "Starting Device Plugin manager" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.638387 1425 manager.go:447] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.638398 1425 server.go:77] "Starting device plugin registration server" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.638796 1425 plugin_watcher.go:52] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.638863 1425 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.638871 1425 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:35.640499 1425 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:35.643172 1425 event.go:267] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.173773346fbc7e2f", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeAllocatableEnforced", Message:"Updated Node Allocatable limit across pods", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 641546287, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 641546287, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'events is forbidden: User "system:anonymous" cannot create resource "events" in API group "" in the namespace "default"' (will not retry!) Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.680071 1425 kubelet_network_linux.go:63] "Initialized iptables rules." protocol=IPv6 Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.680107 1425 status_manager.go:161] "Starting to sync pod status with apiserver" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.680124 1425 kubelet.go:2033] "Starting kubelet main sync loop" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:35.680189 1425 kubelet.go:2057] "Skipping pod synchronization" err="PLEG is not healthy: pleg has yet to be successful" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 15:27:35.682403 1425 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:35.682428 1425 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:35.718913 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.749028 1425 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.750093 1425 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.750127 1425 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.750160 1425 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.750187 1425 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:35.751359 1425 kubelet_node_status.go:94] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:35.751350 1425 event.go:267] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ec0bb0", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientMemory", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientMemory", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 543999408, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 750105646, time.Local), Count:3, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'events "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ec0bb0" is forbidden: User "system:anonymous" cannot patch resource "events" in API group "" in the namespace "default"' (will not retry!) Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:35.752275 1425 event.go:267] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ec7a97", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasNoDiskPressure", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasNoDiskPressure", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 544027799, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 750132863, time.Local), Count:3, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'events "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ec7a97" is forbidden: User "system:anonymous" cannot patch resource "events" in API group "" in the namespace "default"' (will not retry!) Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:35.753346 1425 event.go:267] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ecab0e", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientPID", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientPID", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 544040206, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 750165022, time.Local), Count:3, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'events "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ecab0e" is forbidden: User "system:anonymous" cannot patch resource "events" in API group "" in the namespace "default"' (will not retry!) Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:35.780486 1425 kubelet.go:2119] "SyncLoop ADD" source="file" pods=[] Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:35.819751 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:35.920204 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:36.014699 1425 controller.go:144] failed to ensure lease exists, will retry in 800ms, error: leases.coordination.k8s.io "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" is forbidden: User "system:anonymous" cannot get resource "leases" in API group "coordination.k8s.io" in the namespace "kube-node-lease" Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:36.020790 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:36.121190 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:36.152306 1425 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:36.154692 1425 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:36.154720 1425 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:36.154734 1425 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:36.154760 1425 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:36.156025 1425 kubelet_node_status.go:94] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:36.156018 1425 event.go:267] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ec0bb0", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientMemory", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientMemory", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 543999408, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 15, 27, 36, 154706318, time.Local), Count:4, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'events "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ec0bb0" is forbidden: User "system:anonymous" cannot patch resource "events" in API group "" in the namespace "default"' (will not retry!) Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:36.157004 1425 event.go:267] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ec7a97", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasNoDiskPressure", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasNoDiskPressure", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 544027799, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 15, 27, 36, 154725971, time.Local), Count:4, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'events "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ec7a97" is forbidden: User "system:anonymous" cannot patch resource "events" in API group "" in the namespace "default"' (will not retry!) Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:36.200239 1425 event.go:267] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ecab0e", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientPID", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientPID", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 544040206, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 15, 27, 36, 154738890, time.Local), Count:4, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'events "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ecab0e" is forbidden: User "system:anonymous" cannot patch resource "events" in API group "" in the namespace "default"' (will not retry!) Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 15:27:36.201013 1425 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:36.201037 1425 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:36.221284 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:36.321557 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:36.390043 1425 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 15:27:36.400204 1425 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Node: nodes "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:36.400233 1425 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Node: failed to list *v1.Node: nodes "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:36.401215 1425 certificate_manager.go:270] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2023-01-06 06:46:04 +0000 UTC, rotation deadline is 2023-01-06 03:27:33.041982059 +0000 UTC Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:36.401234 1425 certificate_manager.go:270] kubernetes.io/kube-apiserver-client-kubelet: Waiting 11h59m56.640751334s for next certificate rotation Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:36.421606 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 15:27:36.422692 1425 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:36.422718 1425 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:36.522028 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:36.622427 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:36.722682 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:36.816263 1425 controller.go:144] failed to ensure lease exists, will retry in 1.6s, error: leases.coordination.k8s.io "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" is forbidden: User "system:anonymous" cannot get resource "leases" in API group "coordination.k8s.io" in the namespace "kube-node-lease" Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:36.823338 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 15:27:36.883706 1425 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:36.883735 1425 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:36.924049 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:36.956114 1425 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:36.957258 1425 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:36.957285 1425 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:36.957300 1425 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:36.957531 1425 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:36.959393 1425 event.go:267] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ec0bb0", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientMemory", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientMemory", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 543999408, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 15, 27, 36, 957271801, time.Local), Count:5, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'events "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ec0bb0" is forbidden: User "system:anonymous" cannot patch resource "events" in API group "" in the namespace "default"' (will not retry!) Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:36.959928 1425 kubelet_node_status.go:94] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:36.961003 1425 event.go:267] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ec7a97", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasNoDiskPressure", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasNoDiskPressure", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 544027799, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 15, 27, 36, 957291150, time.Local), Count:5, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'events "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ec7a97" is forbidden: User "system:anonymous" cannot patch resource "events" in API group "" in the namespace "default"' (will not retry!) Jan 05 15:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:36.962485 1425 event.go:267] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ecab0e", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientPID", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientPID", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 544040206, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 15, 27, 36, 957304121, time.Local), Count:5, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'events "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ecab0e" is forbidden: User "system:anonymous" cannot patch resource "events" in API group "" in the namespace "default"' (will not retry!) Jan 05 15:27:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:37.024361 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:37.124722 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:37.225116 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:37.325485 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:37.390050 1425 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Jan 05 15:27:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:37.426562 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:37.526961 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:37.627375 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:37.727810 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:37.828192 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:37.928710 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:38.029105 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:38.129486 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:38.229879 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 15:27:38.240216 1425 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Jan 05 15:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:38.240249 1425 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Jan 05 15:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:38.330489 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:38.389944 1425 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Jan 05 15:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:38.418635 1425 controller.go:144] failed to ensure lease exists, will retry in 3.2s, error: leases.coordination.k8s.io "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" is forbidden: User "system:anonymous" cannot get resource "leases" in API group "coordination.k8s.io" in the namespace "kube-node-lease" Jan 05 15:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:38.430664 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:38.531034 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:38.560169 1425 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 15:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:38.562494 1425 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 15:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:38.562606 1425 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 15:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:38.562686 1425 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 15:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 15:27:38.562762 1425 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 15:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:38.564236 1425 kubelet_node_status.go:94] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 15:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:38.564229 1425 event.go:267] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ec0bb0", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientMemory", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientMemory", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 543999408, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 15, 27, 38, 562588081, time.Local), Count:6, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'events "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ec0bb0" is forbidden: User "system:anonymous" cannot patch resource "events" in API group "" in the namespace "default"' (will not retry!) Jan 05 15:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:38.565266 1425 event.go:267] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ec7a97", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasNoDiskPressure", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasNoDiskPressure", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 544027799, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 15, 27, 38, 562673057, time.Local), Count:6, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'events "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ec7a97" is forbidden: User "system:anonymous" cannot patch resource "events" in API group "" in the namespace "default"' (will not retry!) Jan 05 15:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:38.566259 1425 event.go:267] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ecab0e", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientPID", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientPID", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 544040206, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 15, 27, 38, 562737052, time.Local), Count:6, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'events "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ecab0e" is forbidden: User "system:anonymous" cannot patch resource "events" in API group "" in the namespace "default"' (will not retry!) Jan 05 15:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:38.631609 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 15:27:38.689750 1425 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Jan 05 15:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:38.689776 1425 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Jan 05 15:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 15:27:38.698897 1425 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Node: nodes "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Jan 05 15:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:38.698921 1425 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Node: failed to list *v1.Node: nodes "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Jan 05 15:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 15:27:38.701917 1425 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Jan 05 15:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:38.701938 1425 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Jan 05 15:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:38.732249 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:38.832602 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:38.933097 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:39.033484 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:39.133849 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:39.234244 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 15:27:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 15:27:39.334609 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:40.372386 1425 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Jan 05 07:27:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:40.417727 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:40.518105 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:40.618509 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:40.718906 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:40.819293 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:40.919806 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:41.020188 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:41.120570 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:41.220952 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:41.321450 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:41.371831 1425 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Jan 05 07:27:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:41.421695 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:41.522072 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:41.621869 1425 kubelet.go:2396] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 05 07:27:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:41.622908 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:41.723271 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:41.823655 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:41.924156 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:42.024560 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:42.124908 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:42.225288 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:42.325633 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:42.371999 1425 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Jan 05 07:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:42.426384 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:42.526770 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:42.602253 1425 controller.go:144] failed to ensure lease exists, will retry in 6.4s, error: leases.coordination.k8s.io "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" is forbidden: User "system:anonymous" cannot get resource "leases" in API group "coordination.k8s.io" in the namespace "kube-node-lease" Jan 05 07:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:42.627572 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:42.727821 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:42.746942 1425 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 07:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:42.748194 1425 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 07:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:42.748223 1425 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 07:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:42.748254 1425 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 07:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:42.748281 1425 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 07:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:42.749460 1425 kubelet_node_status.go:94] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 07:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:42.749453 1425 event.go:267] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ec0bb0", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientMemory", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientMemory", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 543999408, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 7, 27, 42, 748209018, time.Local), Count:7, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'events "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ec0bb0" is forbidden: User "system:anonymous" cannot patch resource "events" in API group "" in the namespace "default"' (will not retry!) Jan 05 07:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:42.750464 1425 event.go:267] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ec7a97", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasNoDiskPressure", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasNoDiskPressure", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 544027799, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 7, 27, 42, 748244427, time.Local), Count:7, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'events "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ec7a97" is forbidden: User "system:anonymous" cannot patch resource "events" in API group "" in the namespace "default"' (will not retry!) Jan 05 07:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:42.751346 1425 event.go:267] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ecab0e", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientPID", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientPID", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 15, 27, 35, 544040206, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 7, 27, 42, 748259383, time.Local), Count:7, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'events "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737733469ecab0e" is forbidden: User "system:anonymous" cannot patch resource "events" in API group "" in the namespace "default"' (will not retry!) Jan 05 07:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:42.828867 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:42.929261 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 07:27:42.936465 1425 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Jan 05 07:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:42.936491 1425 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Jan 05 07:27:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:43.029804 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:43.130162 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:43.230533 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:43.331026 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:43.371370 1425 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Jan 05 07:27:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:43.431752 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:43.532132 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:43.632574 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:43.732784 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:43.833283 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:43.933643 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:44.033998 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 07:27:44.123465 1425 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Jan 05 07:27:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:44.123495 1425 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Jan 05 07:27:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:44.134572 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:44.234938 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:44.335439 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:44.372091 1425 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Jan 05 07:27:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:44.436505 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:44.536885 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:44.637384 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 07:27:44.689666 1425 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Node: nodes "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Jan 05 07:27:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:44.689696 1425 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Node: failed to list *v1.Node: nodes "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Jan 05 07:27:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:44.737903 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:44.838273 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:44.938629 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:45.038975 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:45.139336 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 07:27:45.227694 1425 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Jan 05 07:27:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:45.227722 1425 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Jan 05 07:27:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:45.239814 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:45.340259 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:45.371752 1425 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Jan 05 07:27:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:45.441258 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:45.541571 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:45.641910 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:45.742148 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:45.842624 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:45.942981 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:46.043339 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:46.143690 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:46.244059 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:46.316576 1425 transport.go:135] "Certificate rotation detected, shutting down client connections to start using new credentials" Jan 05 07:27:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:46.345049 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:46.384660 1425 nodeinfomanager.go:401] Failed to publish CSINode: nodes "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" not found Jan 05 07:27:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:46.401691 1425 nodeinfomanager.go:401] Failed to publish CSINode: nodes "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" not found Jan 05 07:27:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:46.445926 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:46.459707 1425 nodeinfomanager.go:401] Failed to publish CSINode: nodes "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" not found Jan 05 07:27:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:46.546093 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:46.622716 1425 kubelet.go:2396] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 05 07:27:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:46.622718 1425 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:46.646572 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:46.730854 1425 nodeinfomanager.go:401] Failed to publish CSINode: nodes "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" not found Jan 05 07:27:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:46.730891 1425 csi_plugin.go:295] Failed to initialize CSINode: error updating CSINode annotation: timed out waiting for the condition; caused by: nodes "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" not found Jan 05 07:27:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:46.747106 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:46.753075 1425 nodeinfomanager.go:401] Failed to publish CSINode: nodes "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" not found Jan 05 07:27:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:46.770663 1425 nodeinfomanager.go:401] Failed to publish CSINode: nodes "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" not found Jan 05 07:27:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:46.831861 1425 nodeinfomanager.go:401] Failed to publish CSINode: nodes "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" not found Jan 05 07:27:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:46.847872 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:46.948224 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:47.048573 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:47.149198 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:47.168236 1425 nodeinfomanager.go:401] Failed to publish CSINode: nodes "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" not found Jan 05 07:27:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:47.168268 1425 csi_plugin.go:295] Failed to initialize CSINode: error updating CSINode annotation: timed out waiting for the condition; caused by: nodes "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" not found Jan 05 07:27:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:47.249575 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:47.275522 1425 nodeinfomanager.go:401] Failed to publish CSINode: nodes "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" not found Jan 05 07:27:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:47.292767 1425 nodeinfomanager.go:401] Failed to publish CSINode: nodes "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" not found Jan 05 07:27:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:47.350205 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:47.350313 1425 nodeinfomanager.go:401] Failed to publish CSINode: nodes "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" not found Jan 05 07:27:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:47.450869 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:47.551213 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:47.628590 1425 nodeinfomanager.go:401] Failed to publish CSINode: nodes "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" not found Jan 05 07:27:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:47.628614 1425 csi_plugin.go:295] Failed to initialize CSINode: error updating CSINode annotation: timed out waiting for the condition; caused by: nodes "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" not found Jan 05 07:27:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:47.651392 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:47.751573 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:47.851989 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:47.952168 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:48.053215 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:48.153572 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:48.230309 1425 nodeinfomanager.go:401] Failed to publish CSINode: nodes "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" not found Jan 05 07:27:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:48.248102 1425 nodeinfomanager.go:401] Failed to publish CSINode: nodes "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" not found Jan 05 07:27:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:48.254211 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:48.306991 1425 nodeinfomanager.go:401] Failed to publish CSINode: nodes "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" not found Jan 05 07:27:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:48.354399 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:48.454806 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:48.555166 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:48.572741 1425 nodeinfomanager.go:401] Failed to publish CSINode: nodes "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" not found Jan 05 07:27:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:48.572761 1425 csi_plugin.go:295] Failed to initialize CSINode: error updating CSINode annotation: timed out waiting for the condition; caused by: nodes "pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" not found Jan 05 07:27:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:48.656093 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:48.756265 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:48.856691 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:48.957053 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:49.009923 1425 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 07:27:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:49.057721 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:49.150028 1425 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 07:27:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:49.151207 1425 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 07:27:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:49.151237 1425 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 07:27:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:49.151249 1425 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 07:27:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:49.151279 1425 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 07:27:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:49.158466 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:49.160089 1425 kubelet_node_status.go:75] "Successfully registered node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 07:27:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:49.258939 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:49.359231 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:49.392933 1425 certificate_manager.go:270] kubernetes.io/kubelet-serving: Rotating certificates Jan 05 07:27:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:49.451599 1425 log.go:198] http: TLS handshake error from 10.0.150.214:59692: no serving certificate available for the kubelet Jan 05 07:27:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:49.458638 1425 csr.go:261] certificate signing request csr-n69sh is approved, waiting to be issued Jan 05 07:27:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:49.459656 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:49.475398 1425 csr.go:257] certificate signing request csr-n69sh is issued Jan 05 07:27:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:49.560570 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:49.660987 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:49.761254 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:49.861753 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:49.962021 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:50.062257 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:50.162628 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:50.263031 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:50.363398 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:50.463893 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:50.475967 1425 certificate_manager.go:270] kubernetes.io/kubelet-serving: Certificate expiration is 2023-01-06 06:46:04 +0000 UTC, rotation deadline is 2023-01-06 03:06:47.973709181 +0000 UTC Jan 05 07:27:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:50.475990 1425 certificate_manager.go:270] kubernetes.io/kubelet-serving: Waiting 19h38m57.49772208s for next certificate rotation Jan 05 07:27:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:50.564331 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:50.664922 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:50.765933 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:50.866406 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:50.966749 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:51.067010 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:51.168052 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:51.268442 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:51.369067 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:51.469460 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:51.476584 1425 certificate_manager.go:270] kubernetes.io/kubelet-serving: Certificate expiration is 2023-01-06 06:46:04 +0000 UTC, rotation deadline is 2023-01-06 04:05:32.062430224 +0000 UTC Jan 05 07:27:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:51.476604 1425 certificate_manager.go:270] kubernetes.io/kubelet-serving: Waiting 20h37m40.585829439s for next certificate rotation Jan 05 07:27:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:51.569850 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:51.623468 1425 kubelet.go:2396] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 05 07:27:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:51.669906 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:51.770955 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:51.871350 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:51.971576 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:52.071940 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:52.172179 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:52.272404 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:52.372818 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:52.473056 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:52.574146 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:52.675175 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:52.776123 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:52.876581 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:52.976974 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:53.077357 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:53.177577 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:53.277631 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:53.378189 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:53.478325 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:53.578673 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:53.679293 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:53.779988 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:53.880324 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:53.981357 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:54.082257 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:54.182319 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:54.283333 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:54.383939 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:54.485480 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:54.586243 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:54.686651 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:54.786977 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:54.887255 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:54.987656 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:55.088574 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:55.188911 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:55.288988 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:55.389102 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:55.489518 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:55.589881 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:55.690082 1425 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.371352 1425 apiserver.go:52] "Watching apiserver" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.376702 1425 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-image-registry/node-ca-xw6d2 openshift-machine-config-operator/machine-config-daemon-hf8f5 openshift-network-diagnostics/network-check-target-xmq2g openshift-dns/node-resolver-nxzr8 openshift-cluster-node-tuning-operator/tuned-smwj7 openshift-sdn/sdn-fbccx openshift-multus/multus-twl8f openshift-multus/multus-additional-cni-plugins-5fgxc openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4 openshift-multus/network-metrics-daemon-hq594 openshift-monitoring/node-exporter-2r6nf] Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.376769 1425 topology_manager.go:205] "Topology Admit Handler" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.376842 1425 topology_manager.go:205] "Topology Admit Handler" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.376898 1425 topology_manager.go:205] "Topology Admit Handler" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.376945 1425 topology_manager.go:205] "Topology Admit Handler" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.376985 1425 topology_manager.go:205] "Topology Admit Handler" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.377039 1425 topology_manager.go:205] "Topology Admit Handler" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.377100 1425 topology_manager.go:205] "Topology Admit Handler" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.377157 1425 topology_manager.go:205] "Topology Admit Handler" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.377314 1425 topology_manager.go:205] "Topology Admit Handler" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.377436 1425 topology_manager.go:205] "Topology Admit Handler" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.377559 1425 topology_manager.go:205] "Topology Admit Handler" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:56.378233 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:56.380831 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.478373 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b5d621a5-07d9-459b-8745-a991cf4a179e-env-overrides\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.478696 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-modules\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-modules\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.478737 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-system-cni-dir\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.478774 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-os-release\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.478814 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-textfile\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.478850 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4sjn\" (UniqueName: \"kubernetes.io/projected/659b8bb4-81b5-49bf-8127-1d326f30ba2f-kube-api-access-h4sjn\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.478883 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.478922 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-var-run-dbus\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.478955 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfs2p\" (UniqueName: \"kubernetes.io/projected/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-kube-api-access-dfs2p\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.478987 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-root\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.479018 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-run-crio\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-crio\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.479048 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-cni-bin\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.479082 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7d6lx\" (UniqueName: \"kubernetes.io/projected/b5d621a5-07d9-459b-8745-a991cf4a179e-kube-api-access-7d6lx\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.479111 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-sys\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.479158 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c0e47249-a5ff-460e-958d-53577387c154-cni-binary-copy\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.479206 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e13f5318-440f-4b4e-863d-2489155b9feb-cni-binary-copy\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.479247 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9799x\" (UniqueName: \"kubernetes.io/projected/e13f5318-440f-4b4e-863d-2489155b9feb-kube-api-access-9799x\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.479314 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-kubelet-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.479349 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-lib-modules\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.479382 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-multus-cni-dir\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.479425 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-cnibin\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.479477 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzzxr\" (UniqueName: \"kubernetes.io/projected/1bb84a70-232c-4467-ac0b-647030155908-kube-api-access-vzzxr\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.479503 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5d621a5-07d9-459b-8745-a991cf4a179e-config\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.479532 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-dbus\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.479593 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-rootfs\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.479620 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8f88cbe5-1221-4e6f-a6c9-f22da152b43f-hosts-file\") pod \"node-resolver-nxzr8\" (UID: \"8f88cbe5-1221-4e6f-a6c9-f22da152b43f\") " pod="openshift-dns/node-resolver-nxzr8" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.479655 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/659b8bb4-81b5-49bf-8127-1d326f30ba2f-serviceca\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.479680 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-run-netns\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.479709 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-registration-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.479741 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmxtc\" (UniqueName: \"kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc\") pod \"network-check-target-xmq2g\" (UID: \"f364a949-09d7-4cfb-83ff-e532b822a557\") " pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.479768 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-sys\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.479793 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-host\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.479878 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"non-standard-root-system-trust-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bb84a70-232c-4467-ac0b-647030155908-non-standard-root-system-trust-ca-bundle\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.479931 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-wtmp\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.479961 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-run-ovs\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-ovs\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.480016 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7www\" (UniqueName: \"kubernetes.io/projected/8f88cbe5-1221-4e6f-a6c9-f22da152b43f-kube-api-access-g7www\") pod \"node-resolver-nxzr8\" (UID: \"8f88cbe5-1221-4e6f-a6c9-f22da152b43f\") " pod="openshift-dns/node-resolver-nxzr8" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.480069 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx2gt\" (UniqueName: \"kubernetes.io/projected/50152296-9489-4fa8-aa42-7858debe1c08-kube-api-access-zx2gt\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.480112 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-metrics-client-ca\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.480143 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-mtu-migration\" (UniqueName: \"kubernetes.io/configmap/b5d621a5-07d9-459b-8745-a991cf4a179e-config-mtu-migration\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.480181 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd-system\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-run-systemd-system\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.480221 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-system-cni-dir\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.480253 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-os-release\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.480319 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-lib-cni-networks-openshift-sdn\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.480384 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c0e47249-a5ff-460e-958d-53577387c154-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.480420 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xrks\" (UniqueName: \"kubernetes.io/projected/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-kube-api-access-5xrks\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.480453 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7xz4\" (UniqueName: \"kubernetes.io/projected/91265f01-9dcb-4046-961e-ad4c544224d7-kube-api-access-b7xz4\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.480492 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-plugin-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.480530 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.480635 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-conf\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-cni-conf\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.480693 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-cnibin\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.480736 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wq4k7\" (UniqueName: \"kubernetes.io/projected/c0e47249-a5ff-460e-958d-53577387c154-kube-api-access-wq4k7\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.480778 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-tuned-profiles-data\" (UniqueName: \"kubernetes.io/configmap/91265f01-9dcb-4046-961e-ad4c544224d7-var-lib-tuned-profiles-data\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.480817 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-run-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-kubernetes\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.480861 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/659b8bb4-81b5-49bf-8127-1d326f30ba2f-host\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.480909 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-etc-sysconfig\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.480939 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-etc\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.480992 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-tls\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.481037 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.481072 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-slash\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.481111 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sdn-metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5d621a5-07d9-459b-8745-a991cf4a179e-sdn-metrics-certs\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.481153 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-proxy-tls\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.481193 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cookie-secret\" (UniqueName: \"kubernetes.io/secret/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-cookie-secret\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.481237 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-run-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-openshift-sdn\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.481311 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-device-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.481328 1425 reconciler.go:169] "Reconciler: start to sync state" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:56.514438 1425 watcher.go:152] Failed to watch directory "/sys/fs/cgroup/devices/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0e47249_a5ff_460e_958d_53577387c154.slice": readdirent /sys/fs/cgroup/devices/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0e47249_a5ff_460e_958d_53577387c154.slice: no such file or directory Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.582416 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-etc-sysconfig\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.582598 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"etc\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-etc\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.582638 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-tls\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.582675 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.582707 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-slash\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.582734 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"sdn-metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5d621a5-07d9-459b-8745-a991cf4a179e-sdn-metrics-certs\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.582773 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-proxy-tls\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.582806 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cookie-secret\" (UniqueName: \"kubernetes.io/secret/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-cookie-secret\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.582837 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-run-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-openshift-sdn\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.582869 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-device-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.582901 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b5d621a5-07d9-459b-8745-a991cf4a179e-env-overrides\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.582932 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-modules\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-modules\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.582968 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-system-cni-dir\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.583001 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-os-release\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.583034 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-textfile\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.583077 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-h4sjn\" (UniqueName: \"kubernetes.io/projected/659b8bb4-81b5-49bf-8127-1d326f30ba2f-kube-api-access-h4sjn\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.583111 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.583149 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-var-run-dbus\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.583188 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-dfs2p\" (UniqueName: \"kubernetes.io/projected/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-kube-api-access-dfs2p\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.583221 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-run-crio\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-crio\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.583261 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-cni-bin\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.583326 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-7d6lx\" (UniqueName: \"kubernetes.io/projected/b5d621a5-07d9-459b-8745-a991cf4a179e-kube-api-access-7d6lx\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.583358 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-sys\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.583394 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-root\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.583429 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c0e47249-a5ff-460e-958d-53577387c154-cni-binary-copy\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.583460 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e13f5318-440f-4b4e-863d-2489155b9feb-cni-binary-copy\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.583502 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-9799x\" (UniqueName: \"kubernetes.io/projected/e13f5318-440f-4b4e-863d-2489155b9feb-kube-api-access-9799x\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.583551 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-kubelet-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.583582 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5d621a5-07d9-459b-8745-a991cf4a179e-config\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.583620 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-dbus\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.583662 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-rootfs\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.583695 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8f88cbe5-1221-4e6f-a6c9-f22da152b43f-hosts-file\") pod \"node-resolver-nxzr8\" (UID: \"8f88cbe5-1221-4e6f-a6c9-f22da152b43f\") " pod="openshift-dns/node-resolver-nxzr8" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.583726 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-lib-modules\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.583760 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-multus-cni-dir\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.583794 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-cnibin\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.583829 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-vzzxr\" (UniqueName: \"kubernetes.io/projected/1bb84a70-232c-4467-ac0b-647030155908-kube-api-access-vzzxr\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.583860 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/659b8bb4-81b5-49bf-8127-1d326f30ba2f-serviceca\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.583891 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-run-netns\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.583929 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-registration-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.583966 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-mmxtc\" (UniqueName: \"kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc\") pod \"network-check-target-xmq2g\" (UID: \"f364a949-09d7-4cfb-83ff-e532b822a557\") " pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.583999 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-sys\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.584032 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-host\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.584080 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"non-standard-root-system-trust-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bb84a70-232c-4467-ac0b-647030155908-non-standard-root-system-trust-ca-bundle\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.584116 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-wtmp\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.584146 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-run-ovs\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-ovs\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.584180 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-g7www\" (UniqueName: \"kubernetes.io/projected/8f88cbe5-1221-4e6f-a6c9-f22da152b43f-kube-api-access-g7www\") pod \"node-resolver-nxzr8\" (UID: \"8f88cbe5-1221-4e6f-a6c9-f22da152b43f\") " pod="openshift-dns/node-resolver-nxzr8" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.584214 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-zx2gt\" (UniqueName: \"kubernetes.io/projected/50152296-9489-4fa8-aa42-7858debe1c08-kube-api-access-zx2gt\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.584246 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config-mtu-migration\" (UniqueName: \"kubernetes.io/configmap/b5d621a5-07d9-459b-8745-a991cf4a179e-config-mtu-migration\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.584279 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"run-systemd-system\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-run-systemd-system\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.584334 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-system-cni-dir\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.584365 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-os-release\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.584398 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-metrics-client-ca\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.584438 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-lib-cni-networks-openshift-sdn\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.584479 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c0e47249-a5ff-460e-958d-53577387c154-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.584514 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-5xrks\" (UniqueName: \"kubernetes.io/projected/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-kube-api-access-5xrks\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.584602 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-cni-conf\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-cni-conf\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.584647 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-cnibin\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.584684 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-wq4k7\" (UniqueName: \"kubernetes.io/projected/c0e47249-a5ff-460e-958d-53577387c154-kube-api-access-wq4k7\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.584721 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"var-lib-tuned-profiles-data\" (UniqueName: \"kubernetes.io/configmap/91265f01-9dcb-4046-961e-ad4c544224d7-var-lib-tuned-profiles-data\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.584759 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-b7xz4\" (UniqueName: \"kubernetes.io/projected/91265f01-9dcb-4046-961e-ad4c544224d7-kube-api-access-b7xz4\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.584797 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"plugin-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-plugin-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.584832 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.584865 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-run-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-kubernetes\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.584897 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/659b8bb4-81b5-49bf-8127-1d326f30ba2f-host\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.584985 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/659b8bb4-81b5-49bf-8127-1d326f30ba2f-host\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.584987 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-system-cni-dir\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.582478 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-etc-sysconfig\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.585111 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-os-release\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.585241 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"etc\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-etc\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.585794 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-var-run-dbus\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.585963 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-slash\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.586214 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-run-crio\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-crio\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.586301 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-cni-bin\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.586609 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-sys\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.586683 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"root\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-root\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.586965 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-run-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-openshift-sdn\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.587021 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-device-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.587337 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-kubelet-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.587891 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-dbus\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.588002 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-rootfs\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.588262 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8f88cbe5-1221-4e6f-a6c9-f22da152b43f-hosts-file\") pod \"node-resolver-nxzr8\" (UID: \"8f88cbe5-1221-4e6f-a6c9-f22da152b43f\") " pod="openshift-dns/node-resolver-nxzr8" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.588691 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-modules\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-modules\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.588830 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.589130 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-multus-cni-dir\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.589184 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-cnibin\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.589709 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5d621a5-07d9-459b-8745-a991cf4a179e-config\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.589753 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-run-netns\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.589801 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-registration-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.590030 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-sys\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.590073 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-host\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.590584 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-wtmp\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.590632 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-run-ovs\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-ovs\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.592476 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config-mtu-migration\" (UniqueName: \"kubernetes.io/configmap/b5d621a5-07d9-459b-8745-a991cf4a179e-config-mtu-migration\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.592561 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"run-systemd-system\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-run-systemd-system\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.592617 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-system-cni-dir\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.592670 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-os-release\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.594509 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-metrics-client-ca\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.594590 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-lib-cni-networks-openshift-sdn\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.595933 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"sdn-metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5d621a5-07d9-459b-8745-a991cf4a179e-sdn-metrics-certs\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.597210 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c0e47249-a5ff-460e-958d-53577387c154-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.598059 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-cni-conf\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-cni-conf\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.598143 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-cnibin\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.598611 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c0e47249-a5ff-460e-958d-53577387c154-cni-binary-copy\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:56.599700 1425 secret.go:192] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:56.601042 1425 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs podName:50152296-9489-4fa8-aa42-7858debe1c08 nodeName:}" failed. No retries permitted until 2023-01-05 07:27:57.101022894 +0000 UTC m=+21.088354252 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs") pod "network-metrics-daemon-hq594" (UID: "50152296-9489-4fa8-aa42-7858debe1c08") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.603014 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cookie-secret\" (UniqueName: \"kubernetes.io/secret/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-cookie-secret\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.604805 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-lib-modules\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.605767 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-textfile\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.607732 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/659b8bb4-81b5-49bf-8127-1d326f30ba2f-serviceca\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.608726 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.599759 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-run-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-kubernetes\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.600645 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b5d621a5-07d9-459b-8745-a991cf4a179e-env-overrides\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.609014 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"plugin-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-plugin-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.609056 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e13f5318-440f-4b4e-863d-2489155b9feb-cni-binary-copy\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.610128 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-proxy-tls\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.610419 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"non-standard-root-system-trust-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bb84a70-232c-4467-ac0b-647030155908-non-standard-root-system-trust-ca-bundle\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.610826 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-tls\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.611707 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"var-lib-tuned-profiles-data\" (UniqueName: \"kubernetes.io/configmap/91265f01-9dcb-4046-961e-ad4c544224d7-var-lib-tuned-profiles-data\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.622015 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-7d6lx\" (UniqueName: \"kubernetes.io/projected/b5d621a5-07d9-459b-8745-a991cf4a179e-kube-api-access-7d6lx\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:56.627999 1425 kubelet.go:2396] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.629487 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-9799x\" (UniqueName: \"kubernetes.io/projected/e13f5318-440f-4b4e-863d-2489155b9feb-kube-api-access-9799x\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.631914 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfs2p\" (UniqueName: \"kubernetes.io/projected/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-kube-api-access-dfs2p\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.640660 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzzxr\" (UniqueName: \"kubernetes.io/projected/1bb84a70-232c-4467-ac0b-647030155908-kube-api-access-vzzxr\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.643604 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4sjn\" (UniqueName: \"kubernetes.io/projected/659b8bb4-81b5-49bf-8127-1d326f30ba2f-kube-api-access-h4sjn\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:56.687682 1425 projected.go:290] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:56.687856 1425 projected.go:290] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:56.687952 1425 projected.go:196] Error preparing data for projected volume kube-api-access-mmxtc for pod openshift-network-diagnostics/network-check-target-xmq2g: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:56.688060 1425 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc podName:f364a949-09d7-4cfb-83ff-e532b822a557 nodeName:}" failed. No retries permitted until 2023-01-05 07:27:57.188039314 +0000 UTC m=+21.175370672 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-mmxtc" (UniqueName: "kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc") pod "network-check-target-xmq2g" (UID: "f364a949-09d7-4cfb-83ff-e532b822a557") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.690398 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-wq4k7\" (UniqueName: \"kubernetes.io/projected/c0e47249-a5ff-460e-958d-53577387c154-kube-api-access-wq4k7\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.690895 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx2gt\" (UniqueName: \"kubernetes.io/projected/50152296-9489-4fa8-aa42-7858debe1c08-kube-api-access-zx2gt\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.692396 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7www\" (UniqueName: \"kubernetes.io/projected/8f88cbe5-1221-4e6f-a6c9-f22da152b43f-kube-api-access-g7www\") pod \"node-resolver-nxzr8\" (UID: \"8f88cbe5-1221-4e6f-a6c9-f22da152b43f\") " pod="openshift-dns/node-resolver-nxzr8" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.708031 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xrks\" (UniqueName: \"kubernetes.io/projected/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-kube-api-access-5xrks\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.718920 1425 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:56.719587169Z" level=info msg="Running pod sandbox: openshift-machine-config-operator/machine-config-daemon-hf8f5/POD" id=d06dda77-c0b8-49eb-bc0f-3f776e8bf617 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:56.719855615Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.723741 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7xz4\" (UniqueName: \"kubernetes.io/projected/91265f01-9dcb-4046-961e-ad4c544224d7-kube-api-access-b7xz4\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.730744 1425 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:56.738475588Z" level=info msg="Running pod sandbox: openshift-monitoring/node-exporter-2r6nf/POD" id=829f2cbf-d4d7-4004-8a1c-d3c1abbeb0eb name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:56.738525458Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.744651 1425 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-nxzr8" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:56.746140296Z" level=info msg="Running pod sandbox: openshift-dns/node-resolver-nxzr8/POD" id=500feedd-a2aa-4cf5-92b6-79c50ec41897 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:56.746307758Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.760379 1425 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:56.761461622Z" level=info msg="Running pod sandbox: openshift-cluster-node-tuning-operator/tuned-smwj7/POD" id=172d423b-a390-44c4-9f5e-2d8c027b1cdb name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:56.761624930Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.772262 1425 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-sdn/sdn-fbccx" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:56.772708446Z" level=info msg="Running pod sandbox: openshift-sdn/sdn-fbccx/POD" id=1176ee2a-a669-407d-bf43-80c53911e15c name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:56.772761293Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.789401 1425 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-twl8f" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:56.789772313Z" level=info msg="Running pod sandbox: openshift-multus/multus-twl8f/POD" id=141fbc55-4e24-4d97-bd44-263f1fb07008 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:56.789822628Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.821166 1425 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:56.821876929Z" level=info msg="Running pod sandbox: openshift-multus/multus-additional-cni-plugins-5fgxc/POD" id=b8759fd9-d5c7-440b-955b-8a1b53b4398b name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:56.821918424Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.835132 1425 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-xw6d2" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:56.835571248Z" level=info msg="Running pod sandbox: openshift-image-registry/node-ca-xw6d2/POD" id=682a43d5-ea00-4129-b4c2-f675d0626239 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:56.835627393Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:56.849707 1425 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:56.850062974Z" level=info msg="Running pod sandbox: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/POD" id=51b63e38-221d-4c00-af6d-eb511914ad3b name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:56.850166363Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:57.189525 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-mmxtc\" (UniqueName: \"kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc\") pod \"network-check-target-xmq2g\" (UID: \"f364a949-09d7-4cfb-83ff-e532b822a557\") " pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:57.189593 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:57.189713 1425 secret.go:192] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:57.189771 1425 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs podName:50152296-9489-4fa8-aa42-7858debe1c08 nodeName:}" failed. No retries permitted until 2023-01-05 07:27:58.18975267 +0000 UTC m=+22.177084026 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs") pod "network-metrics-daemon-hq594" (UID: "50152296-9489-4fa8-aa42-7858debe1c08") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:57.189861 1425 projected.go:290] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:57.189895 1425 projected.go:290] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:57.189907 1425 projected.go:196] Error preparing data for projected volume kube-api-access-mmxtc for pod openshift-network-diagnostics/network-check-target-xmq2g: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:57.189979 1425 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc podName:f364a949-09d7-4cfb-83ff-e532b822a557 nodeName:}" failed. No retries permitted until 2023-01-05 07:27:58.189964786 +0000 UTC m=+22.177296137 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-mmxtc" (UniqueName: "kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc") pod "network-check-target-xmq2g" (UID: "f364a949-09d7-4cfb-83ff-e532b822a557") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.345718111Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=d06dda77-c0b8-49eb-bc0f-3f776e8bf617 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.371593806Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=172d423b-a390-44c4-9f5e-2d8c027b1cdb name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.378825794Z" level=info msg="Ran pod sandbox 2cd51e2f6a1cbcdeb26d92c54070ab2a85b957f76038eb175b75e3f2ae639ed7 with infra container: openshift-machine-config-operator/machine-config-daemon-hf8f5/POD" id=d06dda77-c0b8-49eb-bc0f-3f776e8bf617 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.380533052Z" level=info msg="Ran pod sandbox 27b0c41330330b95f91c642446af55e683d47ee10c323ad18406b8c3192efb02 with infra container: openshift-cluster-node-tuning-operator/tuned-smwj7/POD" id=172d423b-a390-44c4-9f5e-2d8c027b1cdb name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.385453858Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ca6ba83315a4882f5c6c4e01c79238e1b794b0e8bfdc2f004990f516e8e0b056" id=013a69fb-bbe2-42aa-a165-b576f72b5c1d name=/runtime.v1.ImageService/ImageStatus Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.385553448Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cb9edf2251410c340a6125ff331924ab57a72ad3166bd9e1305b6d3030284990" id=f10eddae-4446-4e3e-8520-c815e92973cd name=/runtime.v1.ImageService/ImageStatus Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.385847434Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=51b63e38-221d-4c00-af6d-eb511914ad3b name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.386093467Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ca6ba83315a4882f5c6c4e01c79238e1b794b0e8bfdc2f004990f516e8e0b056 not found" id=013a69fb-bbe2-42aa-a165-b576f72b5c1d name=/runtime.v1.ImageService/ImageStatus Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.385926575Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2b208acf1f002c143fdb248dbee2517f7b6d168def4cfef1f3188939123d0dc3,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cb9edf2251410c340a6125ff331924ab57a72ad3166bd9e1305b6d3030284990],Size_:536350945,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=f10eddae-4446-4e3e-8520-c815e92973cd name=/runtime.v1.ImageService/ImageStatus Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.387640287Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cb9edf2251410c340a6125ff331924ab57a72ad3166bd9e1305b6d3030284990" id=fa135133-13b0-438c-9912-a4bc94d5168f name=/runtime.v1.ImageService/ImageStatus Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:57.387911 1425 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.388071748Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2b208acf1f002c143fdb248dbee2517f7b6d168def4cfef1f3188939123d0dc3,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cb9edf2251410c340a6125ff331924ab57a72ad3166bd9e1305b6d3030284990],Size_:536350945,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=fa135133-13b0-438c-9912-a4bc94d5168f name=/runtime.v1.ImageService/ImageStatus Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.390257323Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=1176ee2a-a669-407d-bf43-80c53911e15c name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.396181471Z" level=info msg="Ran pod sandbox 77dc8fbfc80728a344c03166326f04c57592cd6f9e2adc95515af9c2147cc500 with infra container: openshift-sdn/sdn-fbccx/POD" id=1176ee2a-a669-407d-bf43-80c53911e15c name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.398408152Z" level=info msg="Ran pod sandbox d611273f3568d300e64d6a51cb6de91bcf31c3abd27ef78b4fed9dc896457467 with infra container: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/POD" id=51b63e38-221d-4c00-af6d-eb511914ad3b name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.398689274Z" level=info msg="Creating container: openshift-machine-config-operator/machine-config-daemon-hf8f5/machine-config-daemon" id=0d335273-57c6-446a-9c8e-b3c9cb0e59e0 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.398787566Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.398891682Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ca6ba83315a4882f5c6c4e01c79238e1b794b0e8bfdc2f004990f516e8e0b056" id=6bf1bb33-718b-4f3b-b938-024d8434edef name=/runtime.v1.ImageService/PullImage Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.399568417Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=500feedd-a2aa-4cf5-92b6-79c50ec41897 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.407401439Z" level=info msg="Ran pod sandbox 63ad0ecbd8ba6a412b2c14895d6a13d0ead1d3ee31d0f15597f1df586d1d96e9 with infra container: openshift-dns/node-resolver-nxzr8/POD" id=500feedd-a2aa-4cf5-92b6-79c50ec41897 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.407798246Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:30387b6f2e605bb8d018e0c09f9d0275048fc5e7966137fbe6bd4307c88d3fdd" id=0f8515e1-3e18-4404-b4e9-803ad0fdfd91 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.407944315Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7" id=9ba774e7-8cb8-49f7-a601-256b54f2a369 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.408486209Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ca6ba83315a4882f5c6c4e01c79238e1b794b0e8bfdc2f004990f516e8e0b056\"" Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.410149353Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f93a29f8d7455a5067ecde85394835bda877174193ab693b2b010d96c8cd9183" id=7e45bae4-3c0f-4d97-994a-55983c0a4f6f name=/runtime.v1.ImageService/ImageStatus Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.411608075Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7 not found" id=9ba774e7-8cb8-49f7-a601-256b54f2a369 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.411887569Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:30387b6f2e605bb8d018e0c09f9d0275048fc5e7966137fbe6bd4307c88d3fdd not found" id=0f8515e1-3e18-4404-b4e9-803ad0fdfd91 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.412104723Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f93a29f8d7455a5067ecde85394835bda877174193ab693b2b010d96c8cd9183 not found" id=7e45bae4-3c0f-4d97-994a-55983c0a4f6f name=/runtime.v1.ImageService/ImageStatus Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.413344375Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f93a29f8d7455a5067ecde85394835bda877174193ab693b2b010d96c8cd9183" id=974a53cf-31a0-4c12-9c16-3c7d9343ab29 name=/runtime.v1.ImageService/PullImage Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.413610197Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:30387b6f2e605bb8d018e0c09f9d0275048fc5e7966137fbe6bd4307c88d3fdd" id=5eb714ed-abf0-4f4d-a79b-c49c61077f06 name=/runtime.v1.ImageService/PullImage Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.413772352Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7" id=b5cdcd51-e971-4107-8aba-9e1a6f4179b0 name=/runtime.v1.ImageService/PullImage Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.415366375Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7\"" Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.415831461Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f93a29f8d7455a5067ecde85394835bda877174193ab693b2b010d96c8cd9183\"" Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.416140875Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:30387b6f2e605bb8d018e0c09f9d0275048fc5e7966137fbe6bd4307c88d3fdd\"" Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.422005148Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=b8759fd9-d5c7-440b-955b-8a1b53b4398b name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.423929248Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=141fbc55-4e24-4d97-bd44-263f1fb07008 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.425704093Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=682a43d5-ea00-4129-b4c2-f675d0626239 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.427487719Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=829f2cbf-d4d7-4004-8a1c-d3c1abbeb0eb name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.433779712Z" level=info msg="Ran pod sandbox 0b984f430b21595a59efe2294b0b9fce6559367942c6743e20dbab9826dae909 with infra container: openshift-multus/multus-additional-cni-plugins-5fgxc/POD" id=b8759fd9-d5c7-440b-955b-8a1b53b4398b name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.439798813Z" level=info msg="Ran pod sandbox f160f14931d6ef807fb1308f5ac27f062e39c5cad361cedf59e27a46a8b3bd00 with infra container: openshift-multus/multus-twl8f/POD" id=141fbc55-4e24-4d97-bd44-263f1fb07008 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.443581881Z" level=info msg="Ran pod sandbox 7114c6eed49130178fa2f1d156de41d15d9a79d32874a225c7b87798956be18c with infra container: openshift-monitoring/node-exporter-2r6nf/POD" id=829f2cbf-d4d7-4004-8a1c-d3c1abbeb0eb name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.447105953Z" level=info msg="Ran pod sandbox 47c1514985f9106f9edaa2acbae833a41c7f065a534af106d4e033a59d66e2fb with infra container: openshift-image-registry/node-ca-xw6d2/POD" id=682a43d5-ea00-4129-b4c2-f675d0626239 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.448998563Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:3fc008580e4e404960b2de4020a2939ab6421bdf8f51c0b60af4ad903272c741" id=9d6ed5ad-f634-4080-875f-13af82aa75ad name=/runtime.v1.ImageService/ImageStatus Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.449251677Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:3fc008580e4e404960b2de4020a2939ab6421bdf8f51c0b60af4ad903272c741 not found" id=9d6ed5ad-f634-4080-875f-13af82aa75ad name=/runtime.v1.ImageService/ImageStatus Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.461743402Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8" id=9cd1db66-d447-4508-a1df-f811aa5cc9c4 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.462060444Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8 not found" id=9cd1db66-d447-4508-a1df-f811aa5cc9c4 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.468594124Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:3fc008580e4e404960b2de4020a2939ab6421bdf8f51c0b60af4ad903272c741" id=bc6ca33a-c2d4-4724-b410-d3fd107b4491 name=/runtime.v1.ImageService/PullImage Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.468859238Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180" id=bef62e53-d45f-49ff-96ee-8bd8a6530e90 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.469109351Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180 not found" id=bef62e53-d45f-49ff-96ee-8bd8a6530e90 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.469185758Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=a547ef5b-d610-4101-a97f-25c89ad04a09 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.469390082Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f not found" id=a547ef5b-d610-4101-a97f-25c89ad04a09 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.471983652Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:3fc008580e4e404960b2de4020a2939ab6421bdf8f51c0b60af4ad903272c741\"" Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.476866751Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180" id=1448f9fe-5669-445a-a251-9e15bb958114 name=/runtime.v1.ImageService/PullImage Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.477437645Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8" id=923df706-eca8-4da5-b7a3-d6a34e84f4c0 name=/runtime.v1.ImageService/PullImage Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.477613063Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=fdc44dcf-54c9-4782-a70e-b65651339bf8 name=/runtime.v1.ImageService/PullImage Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.487833000Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f\"" Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.488341522Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180\"" Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.488676613Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8\"" Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.531168595Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f93a29f8d7455a5067ecde85394835bda877174193ab693b2b010d96c8cd9183\"" Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.567056549Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180\"" Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.589088331Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7\"" Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.598777290Z" level=info msg="Created container a67a37b160625cdf16ba68c66ec5fec42e012b4abaa1bf82960c626267757950: openshift-machine-config-operator/machine-config-daemon-hf8f5/machine-config-daemon" id=0d335273-57c6-446a-9c8e-b3c9cb0e59e0 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.600024620Z" level=info msg="Starting container: a67a37b160625cdf16ba68c66ec5fec42e012b4abaa1bf82960c626267757950" id=ac21767f-b7e8-409c-ad2e-9d530b1f2c8f name=/runtime.v1.RuntimeService/StartContainer Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.606218534Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:3fc008580e4e404960b2de4020a2939ab6421bdf8f51c0b60af4ad903272c741\"" Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.609982134Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ca6ba83315a4882f5c6c4e01c79238e1b794b0e8bfdc2f004990f516e8e0b056\"" Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.614219938Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:30387b6f2e605bb8d018e0c09f9d0275048fc5e7966137fbe6bd4307c88d3fdd\"" Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.617401815Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f\"" Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.648204104Z" level=info msg="Started container" PID=1576 containerID=a67a37b160625cdf16ba68c66ec5fec42e012b4abaa1bf82960c626267757950 description=openshift-machine-config-operator/machine-config-daemon-hf8f5/machine-config-daemon id=ac21767f-b7e8-409c-ad2e-9d530b1f2c8f name=/runtime.v1.RuntimeService/StartContainer sandboxID=2cd51e2f6a1cbcdeb26d92c54070ab2a85b957f76038eb175b75e3f2ae639ed7 Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.655954966Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8\"" Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:57.662874 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:57.664450 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.683578459Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=d1ca5aa9-e4f9-4375-a9ec-2862e681d064 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.683974010Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b not found" id=d1ca5aa9-e4f9-4375-a9ec-2862e681d064 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.685429793Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=50bad22e-2c38-408b-b3e8-022678576f7f name=/runtime.v1.ImageService/PullImage Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.687305291Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b\"" Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:57.688432 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-nxzr8" event=&{ID:8f88cbe5-1221-4e6f-a6c9-f22da152b43f Type:ContainerStarted Data:63ad0ecbd8ba6a412b2c14895d6a13d0ead1d3ee31d0f15597f1df586d1d96e9} Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:57.689716 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-smwj7" event=&{ID:91265f01-9dcb-4046-961e-ad4c544224d7 Type:ContainerStarted Data:27b0c41330330b95f91c642446af55e683d47ee10c323ad18406b8c3192efb02} Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:57.692600 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-sdn/sdn-fbccx" event=&{ID:b5d621a5-07d9-459b-8745-a991cf4a179e Type:ContainerStarted Data:77dc8fbfc80728a344c03166326f04c57592cd6f9e2adc95515af9c2147cc500} Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:57.699813 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" event=&{ID:1bb84a70-232c-4467-ac0b-647030155908 Type:ContainerStarted Data:d611273f3568d300e64d6a51cb6de91bcf31c3abd27ef78b4fed9dc896457467} Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:57.703488 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" event=&{ID:997b7ab4-939e-465c-9c7d-4a2ebe3a797a Type:ContainerStarted Data:a67a37b160625cdf16ba68c66ec5fec42e012b4abaa1bf82960c626267757950} Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:57.703573 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" event=&{ID:997b7ab4-939e-465c-9c7d-4a2ebe3a797a Type:ContainerStarted Data:2cd51e2f6a1cbcdeb26d92c54070ab2a85b957f76038eb175b75e3f2ae639ed7} Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:57.709206 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-xw6d2" event=&{ID:659b8bb4-81b5-49bf-8127-1d326f30ba2f Type:ContainerStarted Data:47c1514985f9106f9edaa2acbae833a41c7f065a534af106d4e033a59d66e2fb} Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:57.711797 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-twl8f" event=&{ID:e13f5318-440f-4b4e-863d-2489155b9feb Type:ContainerStarted Data:f160f14931d6ef807fb1308f5ac27f062e39c5cad361cedf59e27a46a8b3bd00} Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:57.715205 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerStarted Data:0b984f430b21595a59efe2294b0b9fce6559367942c6743e20dbab9826dae909} Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:57.726454 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-2r6nf" event=&{ID:aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f Type:ContainerStarted Data:7114c6eed49130178fa2f1d156de41d15d9a79d32874a225c7b87798956be18c} Jan 05 07:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:27:57.896058057Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b\"" Jan 05 07:27:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:58.199715 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-mmxtc\" (UniqueName: \"kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc\") pod \"network-check-target-xmq2g\" (UID: \"f364a949-09d7-4cfb-83ff-e532b822a557\") " pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 07:27:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:27:58.199975 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 07:27:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:58.199994 1425 projected.go:290] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 05 07:27:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:58.200026 1425 projected.go:290] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 05 07:27:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:58.200038 1425 projected.go:196] Error preparing data for projected volume kube-api-access-mmxtc for pod openshift-network-diagnostics/network-check-target-xmq2g: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 05 07:27:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:58.200066 1425 secret.go:192] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 05 07:27:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:58.200087 1425 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc podName:f364a949-09d7-4cfb-83ff-e532b822a557 nodeName:}" failed. No retries permitted until 2023-01-05 07:28:00.200071851 +0000 UTC m=+24.187403207 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-mmxtc" (UniqueName: "kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc") pod "network-check-target-xmq2g" (UID: "f364a949-09d7-4cfb-83ff-e532b822a557") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 05 07:27:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:58.200123 1425 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs podName:50152296-9489-4fa8-aa42-7858debe1c08 nodeName:}" failed. No retries permitted until 2023-01-05 07:28:00.20011215 +0000 UTC m=+24.187443492 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs") pod "network-metrics-daemon-hq594" (UID: "50152296-9489-4fa8-aa42-7858debe1c08") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 05 07:27:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:59.663463 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:27:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:27:59.664151 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:28:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:28:00.215188 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-mmxtc\" (UniqueName: \"kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc\") pod \"network-check-target-xmq2g\" (UID: \"f364a949-09d7-4cfb-83ff-e532b822a557\") " pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 07:28:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:28:00.215264 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 07:28:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:00.215501 1425 secret.go:192] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 05 07:28:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:00.215573 1425 projected.go:290] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 05 07:28:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:00.215593 1425 projected.go:290] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 05 07:28:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:00.215606 1425 projected.go:196] Error preparing data for projected volume kube-api-access-mmxtc for pod openshift-network-diagnostics/network-check-target-xmq2g: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 05 07:28:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:00.215655 1425 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs podName:50152296-9489-4fa8-aa42-7858debe1c08 nodeName:}" failed. No retries permitted until 2023-01-05 07:28:04.215633299 +0000 UTC m=+28.202964650 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs") pod "network-metrics-daemon-hq594" (UID: "50152296-9489-4fa8-aa42-7858debe1c08") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 05 07:28:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:00.215731 1425 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc podName:f364a949-09d7-4cfb-83ff-e532b822a557 nodeName:}" failed. No retries permitted until 2023-01-05 07:28:04.215718663 +0000 UTC m=+28.203050009 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-mmxtc" (UniqueName: "kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc") pod "network-check-target-xmq2g" (UID: "f364a949-09d7-4cfb-83ff-e532b822a557") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 05 07:28:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:01.662621 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:28:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:01.663066 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:28:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:02.072127 1425 kubelet.go:2396] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 05 07:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:28:04.263256 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 07:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:28:04.263393 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-mmxtc\" (UniqueName: \"kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc\") pod \"network-check-target-xmq2g\" (UID: \"f364a949-09d7-4cfb-83ff-e532b822a557\") " pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 07:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:04.263527 1425 projected.go:290] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 05 07:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:04.263560 1425 projected.go:290] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 05 07:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:04.263575 1425 projected.go:196] Error preparing data for projected volume kube-api-access-mmxtc for pod openshift-network-diagnostics/network-check-target-xmq2g: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 05 07:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:04.263627 1425 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc podName:f364a949-09d7-4cfb-83ff-e532b822a557 nodeName:}" failed. No retries permitted until 2023-01-05 07:28:12.263608947 +0000 UTC m=+36.250940302 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-mmxtc" (UniqueName: "kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc") pod "network-check-target-xmq2g" (UID: "f364a949-09d7-4cfb-83ff-e532b822a557") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 05 07:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:04.264027 1425 secret.go:192] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 05 07:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:04.264079 1425 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs podName:50152296-9489-4fa8-aa42-7858debe1c08 nodeName:}" failed. No retries permitted until 2023-01-05 07:28:12.264062279 +0000 UTC m=+36.251393623 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs") pod "network-metrics-daemon-hq594" (UID: "50152296-9489-4fa8-aa42-7858debe1c08") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 05 07:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:04.562331 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:04.562865 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:05.913473 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:06.664155 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:28:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:07.073138 1425 kubelet.go:2396] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 05 07:28:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:07.663492 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:28:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:08.664452 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:28:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:09.493017346Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180" id=1448f9fe-5669-445a-a251-9e15bb958114 name=/runtime.v1.ImageService/PullImage Jan 05 07:28:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:09.496599372Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180" id=343d64da-0960-431d-b4f5-684e956197ac name=/runtime.v1.ImageService/ImageStatus Jan 05 07:28:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:09.498899256Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b4cde205857c1e0a984b38c61275a8370136432c7c84dbfbd7f4835f26bbb935,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180],Size_:297268865,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=343d64da-0960-431d-b4f5-684e956197ac name=/runtime.v1.ImageService/ImageStatus Jan 05 07:28:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:09.499717394Z" level=info msg="Creating container: openshift-monitoring/node-exporter-2r6nf/init-textfile" id=4764e623-4b78-435c-b535-3506ff644840 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:28:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:09.499826729Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:28:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:09.613124865Z" level=info msg="Created container 6ed4360ab7858f2258159b4bfa8bcfcdedeedcc7fe6c907152f46fe6b64e2694: openshift-monitoring/node-exporter-2r6nf/init-textfile" id=4764e623-4b78-435c-b535-3506ff644840 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:28:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:09.613760601Z" level=info msg="Starting container: 6ed4360ab7858f2258159b4bfa8bcfcdedeedcc7fe6c907152f46fe6b64e2694" id=2d9ac910-a1eb-48fb-830a-0a910e7aa895 name=/runtime.v1.RuntimeService/StartContainer Jan 05 07:28:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:09.623349292Z" level=info msg="Started container" PID=1790 containerID=6ed4360ab7858f2258159b4bfa8bcfcdedeedcc7fe6c907152f46fe6b64e2694 description=openshift-monitoring/node-exporter-2r6nf/init-textfile id=2d9ac910-a1eb-48fb-830a-0a910e7aa895 name=/runtime.v1.RuntimeService/StartContainer sandboxID=7114c6eed49130178fa2f1d156de41d15d9a79d32874a225c7b87798956be18c Jan 05 07:28:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:09.662911 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:28:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:28:09.936511 1425 generic.go:296] "Generic (PLEG): container finished" podID=aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f containerID="6ed4360ab7858f2258159b4bfa8bcfcdedeedcc7fe6c907152f46fe6b64e2694" exitCode=0 Jan 05 07:28:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:28:09.936572 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-2r6nf" event=&{ID:aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f Type:ContainerDied Data:6ed4360ab7858f2258159b4bfa8bcfcdedeedcc7fe6c907152f46fe6b64e2694} Jan 05 07:28:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:09.937764322Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180" id=c128f6e2-2e2d-458c-a8f7-24405de69779 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:28:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:09.939705270Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b4cde205857c1e0a984b38c61275a8370136432c7c84dbfbd7f4835f26bbb935,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180],Size_:297268865,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=c128f6e2-2e2d-458c-a8f7-24405de69779 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:28:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:09.940582459Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180" id=feecb988-035a-447d-8bfd-acb078d3b9e7 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:28:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:09.942740239Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b4cde205857c1e0a984b38c61275a8370136432c7c84dbfbd7f4835f26bbb935,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180],Size_:297268865,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=feecb988-035a-447d-8bfd-acb078d3b9e7 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:28:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:09.943866473Z" level=info msg="Creating container: openshift-monitoring/node-exporter-2r6nf/node-exporter" id=4970b9e0-f962-4104-abbe-dd4ceb6fcc17 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:28:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:09.944048621Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:10.055883139Z" level=info msg="Created container 89f7c6c064a9b09985310f1dc577ab88864846cb4ea970f919adb7d6196d1049: openshift-monitoring/node-exporter-2r6nf/node-exporter" id=4970b9e0-f962-4104-abbe-dd4ceb6fcc17 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:10.056407499Z" level=info msg="Starting container: 89f7c6c064a9b09985310f1dc577ab88864846cb4ea970f919adb7d6196d1049" id=abc3471f-7c86-4223-8b4c-e5e0d3ab9add name=/runtime.v1.RuntimeService/StartContainer Jan 05 07:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:10.068612310Z" level=info msg="Started container" PID=1907 containerID=89f7c6c064a9b09985310f1dc577ab88864846cb4ea970f919adb7d6196d1049 description=openshift-monitoring/node-exporter-2r6nf/node-exporter id=abc3471f-7c86-4223-8b4c-e5e0d3ab9add name=/runtime.v1.RuntimeService/StartContainer sandboxID=7114c6eed49130178fa2f1d156de41d15d9a79d32874a225c7b87798956be18c Jan 05 07:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:10.084212835Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=d964cf31-f1a2-48fb-a8a0-d1f1cf0a2850 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:10.084498621Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9 not found" id=d964cf31-f1a2-48fb-a8a0-d1f1cf0a2850 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:10.085239435Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=4fb54cb7-d0df-4ef9-b4ac-4614ad9bf14d name=/runtime.v1.ImageService/PullImage Jan 05 07:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:10.086948989Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9\"" Jan 05 07:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:10.222199060Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9\"" Jan 05 07:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:10.663481 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:28:10.939391 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-2r6nf" event=&{ID:aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f Type:ContainerStarted Data:89f7c6c064a9b09985310f1dc577ab88864846cb4ea970f919adb7d6196d1049} Jan 05 07:28:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:11.663508 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:28:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:12.074157 1425 kubelet.go:2396] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 05 07:28:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:28:12.311596 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 07:28:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:28:12.311671 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-mmxtc\" (UniqueName: \"kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc\") pod \"network-check-target-xmq2g\" (UID: \"f364a949-09d7-4cfb-83ff-e532b822a557\") " pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 07:28:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:12.311722 1425 secret.go:192] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 05 07:28:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:12.311770 1425 projected.go:290] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 05 07:28:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:12.311777 1425 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs podName:50152296-9489-4fa8-aa42-7858debe1c08 nodeName:}" failed. No retries permitted until 2023-01-05 07:28:28.311760181 +0000 UTC m=+52.299091531 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs") pod "network-metrics-daemon-hq594" (UID: "50152296-9489-4fa8-aa42-7858debe1c08") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 05 07:28:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:12.311785 1425 projected.go:290] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 05 07:28:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:12.311798 1425 projected.go:196] Error preparing data for projected volume kube-api-access-mmxtc for pod openshift-network-diagnostics/network-check-target-xmq2g: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 05 07:28:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:12.311829 1425 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc podName:f364a949-09d7-4cfb-83ff-e532b822a557 nodeName:}" failed. No retries permitted until 2023-01-05 07:28:28.311817976 +0000 UTC m=+52.299149326 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-mmxtc" (UniqueName: "kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc") pod "network-check-target-xmq2g" (UID: "f364a949-09d7-4cfb-83ff-e532b822a557") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 05 07:28:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:12.663195 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:28:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:13.663006 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:28:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:14.663741 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:28:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:15.662862 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:28:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:16.663791 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:28:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:17.074815 1425 kubelet.go:2396] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 05 07:28:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:17.663053 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:28:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:18.663351 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:28:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:19.663252 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:28:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:20.662751 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:28:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:21.663090 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:28:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:22.075967 1425 kubelet.go:2396] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 05 07:28:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:22.662622 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:28:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:23.662650 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:28:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:24.662720 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:28:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:25.663506 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:28:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:26.519511756Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8" id=923df706-eca8-4da5-b7a3-d6a34e84f4c0 name=/runtime.v1.ImageService/PullImage Jan 05 07:28:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:26.521029803Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8" id=2a8ef393-f646-4bf1-b2e7-1701c3494c5b name=/runtime.v1.ImageService/ImageStatus Jan 05 07:28:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:26.523035609Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:c9e2954d8b9d9e1594d0a16b90086a9297ffd65773a6aa4a64f325e871ce83d1,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8],Size_:363009057,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=2a8ef393-f646-4bf1-b2e7-1701c3494c5b name=/runtime.v1.ImageService/ImageStatus Jan 05 07:28:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:26.523864294Z" level=info msg="Creating container: openshift-image-registry/node-ca-xw6d2/node-ca" id=ee0b564f-4d26-4b5d-8aac-156ebfc28c31 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:28:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:26.524013804Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:28:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:26.665769 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:28:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:26.794124351Z" level=info msg="Created container 53bec495c4494609a435d13a0ed33e8c0dae5ec19be5950e2d960e3a64b6d0b2: openshift-image-registry/node-ca-xw6d2/node-ca" id=ee0b564f-4d26-4b5d-8aac-156ebfc28c31 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:28:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:26.794714456Z" level=info msg="Starting container: 53bec495c4494609a435d13a0ed33e8c0dae5ec19be5950e2d960e3a64b6d0b2" id=5ac5aa63-3c37-4d08-bb6a-7159f767d0a4 name=/runtime.v1.RuntimeService/StartContainer Jan 05 07:28:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:26.805642169Z" level=info msg="Started container" PID=1960 containerID=53bec495c4494609a435d13a0ed33e8c0dae5ec19be5950e2d960e3a64b6d0b2 description=openshift-image-registry/node-ca-xw6d2/node-ca id=5ac5aa63-3c37-4d08-bb6a-7159f767d0a4 name=/runtime.v1.RuntimeService/StartContainer sandboxID=47c1514985f9106f9edaa2acbae833a41c7f065a534af106d4e033a59d66e2fb Jan 05 07:28:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:28:26.982966 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-xw6d2" event=&{ID:659b8bb4-81b5-49bf-8127-1d326f30ba2f Type:ContainerStarted Data:53bec495c4494609a435d13a0ed33e8c0dae5ec19be5950e2d960e3a64b6d0b2} Jan 05 07:28:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:27.077685 1425 kubelet.go:2396] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 05 07:28:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:27.662599 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:28:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:27.746434056Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:3fc008580e4e404960b2de4020a2939ab6421bdf8f51c0b60af4ad903272c741" id=bc6ca33a-c2d4-4724-b410-d3fd107b4491 name=/runtime.v1.ImageService/PullImage Jan 05 07:28:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:27.747840396Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:3fc008580e4e404960b2de4020a2939ab6421bdf8f51c0b60af4ad903272c741" id=77d09d4c-57bd-4213-8c78-ee2298b4afa2 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:28:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:27.749731164Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:6bb2a44a1f0506c18ef471cac7f7e8991d1e6b9d649ccbce50f751dd2863d06d,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:3fc008580e4e404960b2de4020a2939ab6421bdf8f51c0b60af4ad903272c741],Size_:382411122,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=77d09d4c-57bd-4213-8c78-ee2298b4afa2 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:28:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:27.750435312Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/egress-router-binary-copy" id=8a7e87b8-7fd4-42c9-91ed-b8de697bf642 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:28:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:27.750580007Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:28:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:27.969720020Z" level=info msg="Created container 4e3b27757eb5cec1f8e13d4dae0c4070f5a16bbaeef0ea7321a7964f9d9eaf74: openshift-multus/multus-additional-cni-plugins-5fgxc/egress-router-binary-copy" id=8a7e87b8-7fd4-42c9-91ed-b8de697bf642 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:28:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:27.970472796Z" level=info msg="Starting container: 4e3b27757eb5cec1f8e13d4dae0c4070f5a16bbaeef0ea7321a7964f9d9eaf74" id=354adec6-e0b2-4e05-a79b-3bf2f688e736 name=/runtime.v1.RuntimeService/StartContainer Jan 05 07:28:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:27.980092113Z" level=info msg="Started container" PID=2021 containerID=4e3b27757eb5cec1f8e13d4dae0c4070f5a16bbaeef0ea7321a7964f9d9eaf74 description=openshift-multus/multus-additional-cni-plugins-5fgxc/egress-router-binary-copy id=354adec6-e0b2-4e05-a79b-3bf2f688e736 name=/runtime.v1.RuntimeService/StartContainer sandboxID=0b984f430b21595a59efe2294b0b9fce6559367942c6743e20dbab9826dae909 Jan 05 07:28:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:27.989855034Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_f9d88af0-2fb6-4bc5-a112-e6896b96ba23\"" Jan 05 07:28:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:27.990110246Z" level=info msg="Updated default CNI network name to " Jan 05 07:28:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:28.016654073Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/egress-router\"" Jan 05 07:28:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:28.016822896Z" level=info msg="Updated default CNI network name to " Jan 05 07:28:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:28.021378129Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_f9d88af0-2fb6-4bc5-a112-e6896b96ba23\"" Jan 05 07:28:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:28:28.402838 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-mmxtc\" (UniqueName: \"kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc\") pod \"network-check-target-xmq2g\" (UID: \"f364a949-09d7-4cfb-83ff-e532b822a557\") " pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 07:28:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:28:28.402898 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 07:28:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:28.402979 1425 secret.go:192] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 05 07:28:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:28.403037 1425 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs podName:50152296-9489-4fa8-aa42-7858debe1c08 nodeName:}" failed. No retries permitted until 2023-01-05 07:29:00.403020096 +0000 UTC m=+84.390351445 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs") pod "network-metrics-daemon-hq594" (UID: "50152296-9489-4fa8-aa42-7858debe1c08") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 05 07:28:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:28.402981 1425 projected.go:290] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 05 07:28:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:28.403080 1425 projected.go:290] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 05 07:28:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:28.403091 1425 projected.go:196] Error preparing data for projected volume kube-api-access-mmxtc for pod openshift-network-diagnostics/network-check-target-xmq2g: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 05 07:28:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:28.403122 1425 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc podName:f364a949-09d7-4cfb-83ff-e532b822a557 nodeName:}" failed. No retries permitted until 2023-01-05 07:29:00.403112249 +0000 UTC m=+84.390443600 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-mmxtc" (UniqueName: "kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc") pod "network-check-target-xmq2g" (UID: "f364a949-09d7-4cfb-83ff-e532b822a557") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 05 07:28:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:28.665009 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:28:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:28:28.995095 1425 generic.go:296] "Generic (PLEG): container finished" podID=c0e47249-a5ff-460e-958d-53577387c154 containerID="4e3b27757eb5cec1f8e13d4dae0c4070f5a16bbaeef0ea7321a7964f9d9eaf74" exitCode=0 Jan 05 07:28:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:28:28.995147 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerDied Data:4e3b27757eb5cec1f8e13d4dae0c4070f5a16bbaeef0ea7321a7964f9d9eaf74} Jan 05 07:28:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:28.996151402Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d5fc161170fcf68900b2694039803a2e7f1b31eb913566139a41f91851f166d7" id=86c24d15-9bea-4190-b450-2492671cefb1 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:28:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:28.996464030Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d5fc161170fcf68900b2694039803a2e7f1b31eb913566139a41f91851f166d7 not found" id=86c24d15-9bea-4190-b450-2492671cefb1 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:28:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:28.997396067Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d5fc161170fcf68900b2694039803a2e7f1b31eb913566139a41f91851f166d7" id=ce0638ec-3b72-4541-8b73-45810c7d764c name=/runtime.v1.ImageService/PullImage Jan 05 07:28:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:28.998379654Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d5fc161170fcf68900b2694039803a2e7f1b31eb913566139a41f91851f166d7\"" Jan 05 07:28:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:29.663365 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:28:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:30.662844 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:28:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:31.396158893Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d5fc161170fcf68900b2694039803a2e7f1b31eb913566139a41f91851f166d7\"" Jan 05 07:28:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:31.663349 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:28:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:32.078959 1425 kubelet.go:2396] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 05 07:28:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:32.663807 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:28:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:33.662508 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:28:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:34.662932 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:28:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:35.662730 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:28:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:36.663531 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:28:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:37.080057 1425 kubelet.go:2396] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 05 07:28:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:37.663280 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:28:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:38.663805 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:28:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:39.662929 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:28:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:40.663739 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:28:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:41.558260507Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7" id=b5cdcd51-e971-4107-8aba-9e1a6f4179b0 name=/runtime.v1.ImageService/PullImage Jan 05 07:28:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:41.559341120Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7" id=de260b2b-f746-46b5-bf42-18f70628bf4c name=/runtime.v1.ImageService/ImageStatus Jan 05 07:28:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:41.561527524Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:6e78a44a973f2298da6f51c1418094403ee8c6c2b57bde9d3285b9e2ac2ba129,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7],Size_:397889589,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=de260b2b-f746-46b5-bf42-18f70628bf4c name=/runtime.v1.ImageService/ImageStatus Jan 05 07:28:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:41.562468038Z" level=info msg="Creating container: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-driver" id=c6bcdd37-a370-4790-a96b-ab99b5e4c159 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:28:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:41.562593232Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:28:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:41.583316318Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f93a29f8d7455a5067ecde85394835bda877174193ab693b2b010d96c8cd9183" id=974a53cf-31a0-4c12-9c16-3c7d9343ab29 name=/runtime.v1.ImageService/PullImage Jan 05 07:28:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:41.584876014Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f93a29f8d7455a5067ecde85394835bda877174193ab693b2b010d96c8cd9183" id=4c99bb76-fb9c-47ef-91ca-86e956960ffc name=/runtime.v1.ImageService/ImageStatus Jan 05 07:28:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:41.592629469Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:9b1e063a0faa2683ff00ee1f8b324af985c1c142530a63c7f8729ec8dab01b38,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f93a29f8d7455a5067ecde85394835bda877174193ab693b2b010d96c8cd9183],Size_:480731528,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=4c99bb76-fb9c-47ef-91ca-86e956960ffc name=/runtime.v1.ImageService/ImageStatus Jan 05 07:28:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:41.597371456Z" level=info msg="Creating container: openshift-dns/node-resolver-nxzr8/dns-node-resolver" id=5c9ce3c4-9923-410f-ae43-f191f8a34773 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:28:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:41.597472235Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:28:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:41.663269 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:28:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:41.686665575Z" level=info msg="Created container a9e764b1e1e67aa824ae481175ae1cdbb295ecdb79b9e4c75446312f6e7f153e: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-driver" id=c6bcdd37-a370-4790-a96b-ab99b5e4c159 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:28:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:41.687642955Z" level=info msg="Starting container: a9e764b1e1e67aa824ae481175ae1cdbb295ecdb79b9e4c75446312f6e7f153e" id=acf2317c-1574-43eb-8b17-08b5d9dedd35 name=/runtime.v1.RuntimeService/StartContainer Jan 05 07:28:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:41.708972113Z" level=info msg="Started container" PID=2100 containerID=a9e764b1e1e67aa824ae481175ae1cdbb295ecdb79b9e4c75446312f6e7f153e description=openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-driver id=acf2317c-1574-43eb-8b17-08b5d9dedd35 name=/runtime.v1.RuntimeService/StartContainer sandboxID=d611273f3568d300e64d6a51cb6de91bcf31c3abd27ef78b4fed9dc896457467 Jan 05 07:28:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:41.737389286Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:72c764b55efc147864960930ae05b7928c5f976eb76820b38997f0df7afab820" id=62c663c5-4ab0-43f7-b644-7405c51c7ed4 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:28:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:41.737756488Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:72c764b55efc147864960930ae05b7928c5f976eb76820b38997f0df7afab820 not found" id=62c663c5-4ab0-43f7-b644-7405c51c7ed4 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:28:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:41.739276342Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:72c764b55efc147864960930ae05b7928c5f976eb76820b38997f0df7afab820" id=bbeeca1c-a187-4a89-842e-0e720feca137 name=/runtime.v1.ImageService/PullImage Jan 05 07:28:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:41.740854354Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:72c764b55efc147864960930ae05b7928c5f976eb76820b38997f0df7afab820\"" Jan 05 07:28:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:41.944513104Z" level=info msg="Created container 1ca31c50f718f391e5ceffdbce89daf90050b0a6fe1a5fce413ce6c79cec4071: openshift-dns/node-resolver-nxzr8/dns-node-resolver" id=5c9ce3c4-9923-410f-ae43-f191f8a34773 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:28:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:41.945227539Z" level=info msg="Starting container: 1ca31c50f718f391e5ceffdbce89daf90050b0a6fe1a5fce413ce6c79cec4071" id=aa6fa0e9-c318-4f87-b73e-1fde3bfd459e name=/runtime.v1.RuntimeService/StartContainer Jan 05 07:28:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:41.958412645Z" level=info msg="Started container" PID=2140 containerID=1ca31c50f718f391e5ceffdbce89daf90050b0a6fe1a5fce413ce6c79cec4071 description=openshift-dns/node-resolver-nxzr8/dns-node-resolver id=aa6fa0e9-c318-4f87-b73e-1fde3bfd459e name=/runtime.v1.RuntimeService/StartContainer sandboxID=63ad0ecbd8ba6a412b2c14895d6a13d0ead1d3ee31d0f15597f1df586d1d96e9 Jan 05 07:28:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:28:42.027372 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" event=&{ID:1bb84a70-232c-4467-ac0b-647030155908 Type:ContainerStarted Data:a9e764b1e1e67aa824ae481175ae1cdbb295ecdb79b9e4c75446312f6e7f153e} Jan 05 07:28:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:28:42.028508 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-nxzr8" event=&{ID:8f88cbe5-1221-4e6f-a6c9-f22da152b43f Type:ContainerStarted Data:1ca31c50f718f391e5ceffdbce89daf90050b0a6fe1a5fce413ce6c79cec4071} Jan 05 07:28:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:42.081125 1425 kubelet.go:2396] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 05 07:28:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:42.663312 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:28:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:42.901592207Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:72c764b55efc147864960930ae05b7928c5f976eb76820b38997f0df7afab820\"" Jan 05 07:28:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:43.663454 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:28:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:44.662853 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:28:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:45.663231 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:28:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:46.663096 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:28:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:47.083470 1425 kubelet.go:2396] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 05 07:28:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:47.600830905Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=fdc44dcf-54c9-4782-a70e-b65651339bf8 name=/runtime.v1.ImageService/PullImage Jan 05 07:28:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:47.602006560Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=9ffb7aed-5688-4c8e-9597-d1599d53cda6 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:28:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:47.604026917Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=9ffb7aed-5688-4c8e-9597-d1599d53cda6 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:28:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:47.605222962Z" level=info msg="Creating container: openshift-multus/multus-twl8f/kube-multus" id=d8c01cb5-15ea-4847-a410-0bbad081cc1b name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:28:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:47.605345087Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:28:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:47.663467 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:28:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:47.815047172Z" level=info msg="Created container 97ec0f7bd248355e78e3b1b09cc34d015a0c89ef5c3e56dde06e0dea3a0adf60: openshift-multus/multus-twl8f/kube-multus" id=d8c01cb5-15ea-4847-a410-0bbad081cc1b name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:28:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:47.815636751Z" level=info msg="Starting container: 97ec0f7bd248355e78e3b1b09cc34d015a0c89ef5c3e56dde06e0dea3a0adf60" id=5e7e0107-926c-4824-a0bd-3ae62a443082 name=/runtime.v1.RuntimeService/StartContainer Jan 05 07:28:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:47.825278610Z" level=info msg="Started container" PID=2188 containerID=97ec0f7bd248355e78e3b1b09cc34d015a0c89ef5c3e56dde06e0dea3a0adf60 description=openshift-multus/multus-twl8f/kube-multus id=5e7e0107-926c-4824-a0bd-3ae62a443082 name=/runtime.v1.RuntimeService/StartContainer sandboxID=f160f14931d6ef807fb1308f5ac27f062e39c5cad361cedf59e27a46a8b3bd00 Jan 05 07:28:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:47.832122486Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_4886a84b-3085-493a-b3d6-15a19d587d2a\"" Jan 05 07:28:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:47.839039690Z" level=info msg="Updated default CNI network name to " Jan 05 07:28:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:47.868491396Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/multus\"" Jan 05 07:28:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:47.868567298Z" level=info msg="Updated default CNI network name to " Jan 05 07:28:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:47.873587634Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_4886a84b-3085-493a-b3d6-15a19d587d2a\"" Jan 05 07:28:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:47.976622736Z" level=info msg="CNI monitoring event CREATE \"/etc/kubernetes/cni/net.d/multus.d\"" Jan 05 07:28:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:47.976705022Z" level=info msg="Updated default CNI network name to " Jan 05 07:28:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:28:48.044646 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-twl8f" event=&{ID:e13f5318-440f-4b4e-863d-2489155b9feb Type:ContainerStarted Data:97ec0f7bd248355e78e3b1b09cc34d015a0c89ef5c3e56dde06e0dea3a0adf60} Jan 05 07:28:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:48.663440 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:28:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:49.662669 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:50.662764 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:28:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:51.663445 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:28:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:52.084175 1425 kubelet.go:2396] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 05 07:28:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:52.663408 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:28:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:53.663266 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:28:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:54.664734 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:28:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:54.780705584Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:72c764b55efc147864960930ae05b7928c5f976eb76820b38997f0df7afab820" id=bbeeca1c-a187-4a89-842e-0e720feca137 name=/runtime.v1.ImageService/PullImage Jan 05 07:28:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:54.783126106Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:72c764b55efc147864960930ae05b7928c5f976eb76820b38997f0df7afab820" id=ac15771b-d02e-4544-a53e-714f7e243a5f name=/runtime.v1.ImageService/ImageStatus Jan 05 07:28:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:54.784882474Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:0bb939dc4e915f972f52c355cea411db098a9c605b2ca4bc1048447ae297391e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:72c764b55efc147864960930ae05b7928c5f976eb76820b38997f0df7afab820],Size_:293414029,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=ac15771b-d02e-4544-a53e-714f7e243a5f name=/runtime.v1.ImageService/ImageStatus Jan 05 07:28:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:54.786155759Z" level=info msg="Creating container: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-node-driver-registrar" id=efb4ee93-42a6-4cea-86ca-4a132be5c972 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:28:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:54.786258935Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:28:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:54.895497445Z" level=info msg="Created container be04db214b78889d0ac45006c4180a5d52d68db3eb83038957ee447885fe30d2: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-node-driver-registrar" id=efb4ee93-42a6-4cea-86ca-4a132be5c972 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:28:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:54.896146496Z" level=info msg="Starting container: be04db214b78889d0ac45006c4180a5d52d68db3eb83038957ee447885fe30d2" id=8fbe3d27-77f8-44c2-85f9-8e77dd983b8c name=/runtime.v1.RuntimeService/StartContainer Jan 05 07:28:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:54.906188064Z" level=info msg="Started container" PID=2373 containerID=be04db214b78889d0ac45006c4180a5d52d68db3eb83038957ee447885fe30d2 description=openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-node-driver-registrar id=8fbe3d27-77f8-44c2-85f9-8e77dd983b8c name=/runtime.v1.RuntimeService/StartContainer sandboxID=d611273f3568d300e64d6a51cb6de91bcf31c3abd27ef78b4fed9dc896457467 Jan 05 07:28:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:54.922586187Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ea25c2f3749dee6981359b4655a9acfcd4d95547066ccdf328354b548da74932" id=3dcfc84b-0baa-45e0-9cc1-ca2f233d6793 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:28:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:54.923014788Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ea25c2f3749dee6981359b4655a9acfcd4d95547066ccdf328354b548da74932 not found" id=3dcfc84b-0baa-45e0-9cc1-ca2f233d6793 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:28:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:54.924772915Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ea25c2f3749dee6981359b4655a9acfcd4d95547066ccdf328354b548da74932" id=fc093b54-00b4-4733-88b1-6c70aad508db name=/runtime.v1.ImageService/PullImage Jan 05 07:28:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:54.926528288Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ea25c2f3749dee6981359b4655a9acfcd4d95547066ccdf328354b548da74932\"" Jan 05 07:28:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:28:55.060997 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" event=&{ID:1bb84a70-232c-4467-ac0b-647030155908 Type:ContainerStarted Data:be04db214b78889d0ac45006c4180a5d52d68db3eb83038957ee447885fe30d2} Jan 05 07:28:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:55.663344 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:28:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:28:55.713603077Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ea25c2f3749dee6981359b4655a9acfcd4d95547066ccdf328354b548da74932\"" Jan 05 07:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:56.665336 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:57.099144 1425 kubelet.go:2396] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 05 07:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:57.663264 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:28:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:58.664753 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:28:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:28:59.663522 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:29:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:00.412555 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-mmxtc\" (UniqueName: \"kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc\") pod \"network-check-target-xmq2g\" (UID: \"f364a949-09d7-4cfb-83ff-e532b822a557\") " pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 07:29:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:00.412619 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 07:29:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:29:00.412737 1425 secret.go:192] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 05 07:29:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:29:00.412797 1425 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs podName:50152296-9489-4fa8-aa42-7858debe1c08 nodeName:}" failed. No retries permitted until 2023-01-05 07:30:04.412778185 +0000 UTC m=+148.400109538 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs") pod "network-metrics-daemon-hq594" (UID: "50152296-9489-4fa8-aa42-7858debe1c08") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 05 07:29:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:29:00.412892 1425 projected.go:290] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 05 07:29:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:29:00.412911 1425 projected.go:290] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 05 07:29:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:29:00.412923 1425 projected.go:196] Error preparing data for projected volume kube-api-access-mmxtc for pod openshift-network-diagnostics/network-check-target-xmq2g: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 05 07:29:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:29:00.412961 1425 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc podName:f364a949-09d7-4cfb-83ff-e532b822a557 nodeName:}" failed. No retries permitted until 2023-01-05 07:30:04.412946781 +0000 UTC m=+148.400278136 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-mmxtc" (UniqueName: "kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc") pod "network-check-target-xmq2g" (UID: "f364a949-09d7-4cfb-83ff-e532b822a557") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 05 07:29:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:29:00.663242 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:29:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:29:01.663360 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:29:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:29:02.100593 1425 kubelet.go:2396] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 05 07:29:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:29:02.664854 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:29:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:29:03.662508 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:29:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:29:04.663746 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.420713426Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ea25c2f3749dee6981359b4655a9acfcd4d95547066ccdf328354b548da74932" id=fc093b54-00b4-4733-88b1-6c70aad508db name=/runtime.v1.ImageService/PullImage Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.422034480Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ea25c2f3749dee6981359b4655a9acfcd4d95547066ccdf328354b548da74932" id=2ec9c2bb-17b1-48fb-afc6-da1415603a19 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.424442152Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:58a5591f9334b70902e27f9a172f3e141b8f07a879507b850a76c2499014eec9,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ea25c2f3749dee6981359b4655a9acfcd4d95547066ccdf328354b548da74932],Size_:291396124,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=2ec9c2bb-17b1-48fb-afc6-da1415603a19 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.426118243Z" level=info msg="Creating container: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-liveness-probe" id=d7f1f47c-fcaf-49b0-9481-c5002ac57e81 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.426243185Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.457134287Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:30387b6f2e605bb8d018e0c09f9d0275048fc5e7966137fbe6bd4307c88d3fdd" id=5eb714ed-abf0-4f4d-a79b-c49c61077f06 name=/runtime.v1.ImageService/PullImage Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.469644389Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:30387b6f2e605bb8d018e0c09f9d0275048fc5e7966137fbe6bd4307c88d3fdd" id=22a43f44-0b1c-4342-83c6-571eccba16f4 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.475828977Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:cff39d57d2c205a584f8bf4299b1e125fa1a96ac92f680117ba82c20334754d1,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:30387b6f2e605bb8d018e0c09f9d0275048fc5e7966137fbe6bd4307c88d3fdd],Size_:701241664,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=22a43f44-0b1c-4342-83c6-571eccba16f4 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.486750259Z" level=info msg="Creating container: openshift-sdn/sdn-fbccx/sdn" id=11c4b0df-3e72-4206-bfb8-a0182ab5302f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.486868594Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.599096804Z" level=info msg="Created container 41f5998c3c51852cac932c9963cabdd75a3ea17db16a2164c474aedce445d0da: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-liveness-probe" id=d7f1f47c-fcaf-49b0-9481-c5002ac57e81 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.600268737Z" level=info msg="Starting container: 41f5998c3c51852cac932c9963cabdd75a3ea17db16a2164c474aedce445d0da" id=c8b391b4-65c7-440a-b6b3-defb581be6c1 name=/runtime.v1.RuntimeService/StartContainer Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.621976840Z" level=info msg="Started container" PID=2525 containerID=41f5998c3c51852cac932c9963cabdd75a3ea17db16a2164c474aedce445d0da description=openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-liveness-probe id=c8b391b4-65c7-440a-b6b3-defb581be6c1 name=/runtime.v1.RuntimeService/StartContainer sandboxID=d611273f3568d300e64d6a51cb6de91bcf31c3abd27ef78b4fed9dc896457467 Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.641245815Z" level=info msg="Created container 56f8ab4efdd91320957f5f42e237237e6768ff86ea26e753e3315a2ebb62fdf0: openshift-sdn/sdn-fbccx/sdn" id=11c4b0df-3e72-4206-bfb8-a0182ab5302f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.641728209Z" level=info msg="Starting container: 56f8ab4efdd91320957f5f42e237237e6768ff86ea26e753e3315a2ebb62fdf0" id=1ff26c62-33d4-44e3-9a77-7fd74268b2cc name=/runtime.v1.RuntimeService/StartContainer Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.651158944Z" level=info msg="Started container" PID=2548 containerID=56f8ab4efdd91320957f5f42e237237e6768ff86ea26e753e3315a2ebb62fdf0 description=openshift-sdn/sdn-fbccx/sdn id=1ff26c62-33d4-44e3-9a77-7fd74268b2cc name=/runtime.v1.RuntimeService/StartContainer sandboxID=77dc8fbfc80728a344c03166326f04c57592cd6f9e2adc95515af9c2147cc500 Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.662486374Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.662694109Z" level=info msg="Updated default CNI network name to " Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.662773894Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.662828364Z" level=info msg="Updated default CNI network name to " Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:29:05.663794 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.664221802Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.664346284Z" level=info msg="Updated default CNI network name to " Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.664881241Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.664962570Z" level=info msg="Updated default CNI network name to " Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.664990299Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.665017451Z" level=info msg="Updated default CNI network name to " Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.665062801Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.665092608Z" level=info msg="Updated default CNI network name to " Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.665201112Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.665267143Z" level=info msg="Updated default CNI network name to " Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.665315782Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.665345238Z" level=info msg="Updated default CNI network name to " Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.665364464Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.665390997Z" level=info msg="Updated default CNI network name to " Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.665409250Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.665435747Z" level=info msg="Updated default CNI network name to " Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.665969951Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.666080464Z" level=info msg="Updated default CNI network name to " Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.667628465Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.667775099Z" level=info msg="Updated default CNI network name to " Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.670136961Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.671179812Z" level=info msg="Updated default CNI network name to " Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.671414549Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.671565535Z" level=info msg="Updated default CNI network name to " Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.671651231Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.671742060Z" level=info msg="Updated default CNI network name to " Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.671819476Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.671931995Z" level=info msg="Updated default CNI network name to " Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.672604223Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.673757414Z" level=info msg="Updated default CNI network name to " Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.673988003Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.674128681Z" level=info msg="Updated default CNI network name to " Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.674285599Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=cb22e3ae-0b7f-4e8e-b6b8-745ba0c7d1ae name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.675133644Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9 not found" id=cb22e3ae-0b7f-4e8e-b6b8-745ba0c7d1ae name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.676020654Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.677823952Z" level=info msg="Updated default CNI network name to " Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.678259665Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.678415575Z" level=info msg="Updated default CNI network name to " Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.678492012Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.679268481Z" level=info msg="Updated default CNI network name to " Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.676841339Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=ca380ba9-ccf3-42b9-a431-dd14eaa3b18a name=/runtime.v1.ImageService/PullImage Jan 05 07:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:05.681210712Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9\"" Jan 05 07:29:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:06.036983844Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9\"" Jan 05 07:29:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:06.386848 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-sdn/sdn-fbccx" event=&{ID:b5d621a5-07d9-459b-8745-a991cf4a179e Type:ContainerStarted Data:56f8ab4efdd91320957f5f42e237237e6768ff86ea26e753e3315a2ebb62fdf0} Jan 05 07:29:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:06.389359 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" event=&{ID:1bb84a70-232c-4467-ac0b-647030155908 Type:ContainerStarted Data:41f5998c3c51852cac932c9963cabdd75a3ea17db16a2164c474aedce445d0da} Jan 05 07:29:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:29:06.664393 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:29:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:06.851538 1425 patch_prober.go:29] interesting pod/alibaba-disk-csi-driver-node-5sqb4 container/csi-driver namespace/openshift-cluster-csi-drivers: Liveness probe status=failure output="Get \"http://10.0.99.113:10300/healthz\": dial tcp 10.0.99.113:10300: connect: connection refused" start-of-body= Jan 05 07:29:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:06.851641 1425 prober.go:114] "Probe failed" probeType="Liveness" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" podUID=1bb84a70-232c-4467-ac0b-647030155908 containerName="csi-driver" probeResult=failure output="Get \"http://10.0.99.113:10300/healthz\": dial tcp 10.0.99.113:10300: connect: connection refused" Jan 05 07:29:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:29:07.130962 1425 kubelet.go:2396] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 05 07:29:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:29:07.663178 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:29:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:08.346453379Z" level=info msg="CNI monitoring event CREATE \"/etc/kubernetes/cni/net.d/00-multus.conf\"" Jan 05 07:29:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:08.367498658Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 07:29:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:08.367553080Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 07:29:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:08.367571380Z" level=info msg="CNI monitoring event WRITE \"/etc/kubernetes/cni/net.d/00-multus.conf\"" Jan 05 07:29:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:08.380690119Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 07:29:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:08.380725994Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 07:29:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:08.380741394Z" level=info msg="CNI monitoring event CHMOD \"/etc/kubernetes/cni/net.d/00-multus.conf\"" Jan 05 07:29:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:29:08.663377 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:29:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:29:09.662754 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:29:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:29:10.663044 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 07:29:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 07:29:11.662488 1425 pod_workers.go:965] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: No CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 07:29:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:11.882582 1425 plugin_watcher.go:203] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/diskplugin.csi.alibabacloud.com-reg.sock" Jan 05 07:29:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:12.683141 1425 reconciler.go:164] "OperationExecutor.RegisterPlugin started" plugin={SocketPath:/var/lib/kubelet/plugins_registry/diskplugin.csi.alibabacloud.com-reg.sock Timestamp:2023-01-05 07:29:11.882605385 +0000 UTC m=+95.869936735 Handler: Name:} Jan 05 07:29:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:12.685884 1425 csi_plugin.go:99] kubernetes.io/csi: Trying to validate a new CSI Driver with name: diskplugin.csi.alibabacloud.com endpoint: /var/lib/kubelet/plugins/diskplugin.csi.alibabacloud.com/csi.sock versions: 1.0.0 Jan 05 07:29:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:12.685911 1425 csi_plugin.go:112] kubernetes.io/csi: Register new plugin with name: diskplugin.csi.alibabacloud.com at endpoint: /var/lib/kubelet/plugins/diskplugin.csi.alibabacloud.com/csi.sock Jan 05 07:29:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:17.992500060Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=4fb54cb7-d0df-4ef9-b4ac-4614ad9bf14d name=/runtime.v1.ImageService/PullImage Jan 05 07:29:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:17.994065500Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=7328dacf-557d-494e-bb1a-df0bdc05a5a8 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:17.995846001Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=7328dacf-557d-494e-bb1a-df0bdc05a5a8 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:17.996897646Z" level=info msg="Creating container: openshift-monitoring/node-exporter-2r6nf/kube-rbac-proxy" id=f952db23-cdc5-4742-94c0-bf03c0a54e0b name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:29:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:17.997002340Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:29:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:18.161011569Z" level=info msg="Created container 0ce2612c11b0ce32170c53af3bda82f04dc2d8fdac6efef08145606b9a264cb4: openshift-monitoring/node-exporter-2r6nf/kube-rbac-proxy" id=f952db23-cdc5-4742-94c0-bf03c0a54e0b name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:29:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:18.164921437Z" level=info msg="Starting container: 0ce2612c11b0ce32170c53af3bda82f04dc2d8fdac6efef08145606b9a264cb4" id=c896f584-cba8-4e02-8f42-28949d27e37c name=/runtime.v1.RuntimeService/StartContainer Jan 05 07:29:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:18.185146568Z" level=info msg="Started container" PID=3015 containerID=0ce2612c11b0ce32170c53af3bda82f04dc2d8fdac6efef08145606b9a264cb4 description=openshift-monitoring/node-exporter-2r6nf/kube-rbac-proxy id=c896f584-cba8-4e02-8f42-28949d27e37c name=/runtime.v1.RuntimeService/StartContainer sandboxID=7114c6eed49130178fa2f1d156de41d15d9a79d32874a225c7b87798956be18c Jan 05 07:29:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:18.485450 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-2r6nf" event=&{ID:aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f Type:ContainerStarted Data:0ce2612c11b0ce32170c53af3bda82f04dc2d8fdac6efef08145606b9a264cb4} Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:21.365225 1425 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeReady" Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:21.419814 1425 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-dns/dns-default-2pfzf] Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:21.419871 1425 topology_manager.go:205] "Topology Admit Handler" Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:21.448428 1425 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-ingress-canary/ingress-canary-8xrbh] Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:21.448467 1425 topology_manager.go:205] "Topology Admit Handler" Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:21.506905 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-config-volume\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:21.507135 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzfn5\" (UniqueName: \"kubernetes.io/projected/e04a47b4-aa8e-4d4e-9d45-aa8037dcb748-kube-api-access-lzfn5\") pod \"ingress-canary-8xrbh\" (UID: \"e04a47b4-aa8e-4d4e-9d45-aa8037dcb748\") " pod="openshift-ingress-canary/ingress-canary-8xrbh" Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:21.507797 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-metrics-tls\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:21.508000 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g48vj\" (UniqueName: \"kubernetes.io/projected/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-kube-api-access-g48vj\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:21.609750 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-metrics-tls\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:21.621982 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-g48vj\" (UniqueName: \"kubernetes.io/projected/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-kube-api-access-g48vj\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:21.634724 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-config-volume\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:21.638631 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-lzfn5\" (UniqueName: \"kubernetes.io/projected/e04a47b4-aa8e-4d4e-9d45-aa8037dcb748-kube-api-access-lzfn5\") pod \"ingress-canary-8xrbh\" (UID: \"e04a47b4-aa8e-4d4e-9d45-aa8037dcb748\") " pod="openshift-ingress-canary/ingress-canary-8xrbh" Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:21.640105 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-config-volume\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:21.640695 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-metrics-tls\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:21.669395 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzfn5\" (UniqueName: \"kubernetes.io/projected/e04a47b4-aa8e-4d4e-9d45-aa8037dcb748-kube-api-access-lzfn5\") pod \"ingress-canary-8xrbh\" (UID: \"e04a47b4-aa8e-4d4e-9d45-aa8037dcb748\") " pod="openshift-ingress-canary/ingress-canary-8xrbh" Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:21.675423 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-g48vj\" (UniqueName: \"kubernetes.io/projected/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-kube-api-access-g48vj\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:21.775974732Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ca6ba83315a4882f5c6c4e01c79238e1b794b0e8bfdc2f004990f516e8e0b056" id=6bf1bb33-718b-4f3b-b938-024d8434edef name=/runtime.v1.ImageService/PullImage Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:21.801168 1425 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-8xrbh" Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:21.801651131Z" level=info msg="Running pod sandbox: openshift-ingress-canary/ingress-canary-8xrbh/POD" id=5d12a229-af97-4093-8ea9-4f5e569eadf6 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:21.801849504Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:21.803821961Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ca6ba83315a4882f5c6c4e01c79238e1b794b0e8bfdc2f004990f516e8e0b056" id=e6cb893a-d059-4312-89b3-59c7259e24c5 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:21.808088 1425 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-2pfzf" Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:21.808524061Z" level=info msg="Running pod sandbox: openshift-dns/dns-default-2pfzf/POD" id=00814159-282e-4ca6-86b0-cb7cabd608b3 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:21.808586858Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:21.814215672Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b2f8382287975a85853941eec2c606723c2ed1c24f59fa46ebdb470b91a481f5,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ca6ba83315a4882f5c6c4e01c79238e1b794b0e8bfdc2f004990f516e8e0b056],Size_:505526574,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=e6cb893a-d059-4312-89b3-59c7259e24c5 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:21.815648295Z" level=info msg="Creating container: openshift-cluster-node-tuning-operator/tuned-smwj7/tuned" id=b834d3f0-dc91-4bef-98df-37d0cf6172f9 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:21.815745583Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:21.824209270Z" level=info msg="Got pod network &{Name:ingress-canary-8xrbh Namespace:openshift-ingress-canary ID:3e1262384ce877333f952ac7ba12d68b4133181f637b84095f42300e1641af9e UID:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748 NetNS:/var/run/netns/91970716-b3de-4e92-b724-0e63fbf7eb3a Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:21.824371482Z" level=info msg="Adding pod openshift-ingress-canary_ingress-canary-8xrbh to CNI network \"multus-cni-network\" (type=multus)" Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:21.842017439Z" level=info msg="Got pod network &{Name:dns-default-2pfzf Namespace:openshift-dns ID:81b981c00eed3ce8e7b961151a721e4be9ef2eef83f4433fad233b19e63776bf UID:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 NetNS:/var/run/netns/7b5530ed-9b90-41d9-b3c4-0a53de335ef7 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:21.842061731Z" level=info msg="Adding pod openshift-dns_dns-default-2pfzf to CNI network \"multus-cni-network\" (type=multus)" Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:21.973590707Z" level=info msg="Created container eec0587c6ad94c72821fb4fc320862719d46352e5f9690b166cc6ff42529a280: openshift-cluster-node-tuning-operator/tuned-smwj7/tuned" id=b834d3f0-dc91-4bef-98df-37d0cf6172f9 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:29:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:21.974740046Z" level=info msg="Starting container: eec0587c6ad94c72821fb4fc320862719d46352e5f9690b166cc6ff42529a280" id=acea9219-9c06-413c-9d06-aaad1718fd73 name=/runtime.v1.RuntimeService/StartContainer Jan 05 07:29:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:22.057484919Z" level=info msg="Started container" PID=3144 containerID=eec0587c6ad94c72821fb4fc320862719d46352e5f9690b166cc6ff42529a280 description=openshift-cluster-node-tuning-operator/tuned-smwj7/tuned id=acea9219-9c06-413c-9d06-aaad1718fd73 name=/runtime.v1.RuntimeService/StartContainer sandboxID=27b0c41330330b95f91c642446af55e683d47ee10c323ad18406b8c3192efb02 Jan 05 07:29:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:22.517611 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-smwj7" event=&{ID:91265f01-9dcb-4046-961e-ad4c544224d7 Type:ContainerStarted Data:eec0587c6ad94c72821fb4fc320862719d46352e5f9690b166cc6ff42529a280} Jan 05 07:29:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: 2023-01-05T07:29:23Z [verbose] Add: openshift-ingress-canary:ingress-canary-8xrbh:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/91970716-b3de-4e92-b724-0e63fbf7eb3a"}],"ips":[{"version":"4","interface":0,"address":"10.129.2.3/23","gateway":"10.129.2.1"}],"dns":{}} Jan 05 07:29:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: I0105 07:29:23.874585 3119 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-ingress-canary", Name:"ingress-canary-8xrbh", UID:"e04a47b4-aa8e-4d4e-9d45-aa8037dcb748", APIVersion:"v1", ResourceVersion:"35730", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.129.2.3/23] from openshift-sdn Jan 05 07:29:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:23.888500701Z" level=info msg="Got pod network &{Name:ingress-canary-8xrbh Namespace:openshift-ingress-canary ID:3e1262384ce877333f952ac7ba12d68b4133181f637b84095f42300e1641af9e UID:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748 NetNS:/var/run/netns/91970716-b3de-4e92-b724-0e63fbf7eb3a Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 07:29:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:23.888664254Z" level=info msg="Checking pod openshift-ingress-canary_ingress-canary-8xrbh for CNI network multus-cni-network (type=multus)" Jan 05 07:29:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:23.890746 1425 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-ingress-canary/ingress-canary-8xrbh] Jan 05 07:29:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 07:29:23.897923 1425 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode04a47b4_aa8e_4d4e_9d45_aa8037dcb748.slice/crio-3e1262384ce877333f952ac7ba12d68b4133181f637b84095f42300e1641af9e.scope WatchSource:0}: Error finding container 3e1262384ce877333f952ac7ba12d68b4133181f637b84095f42300e1641af9e: Status 404 returned error can't find the container with id 3e1262384ce877333f952ac7ba12d68b4133181f637b84095f42300e1641af9e Jan 05 07:29:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:23.901123986Z" level=info msg="Ran pod sandbox 3e1262384ce877333f952ac7ba12d68b4133181f637b84095f42300e1641af9e with infra container: openshift-ingress-canary/ingress-canary-8xrbh/POD" id=5d12a229-af97-4093-8ea9-4f5e569eadf6 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:29:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:23.903637534Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:5974989cda2655deb3052d601e7406eaf3caee0e7999a6eff251760aac19c199" id=4b0a74f3-6a59-457b-9fa1-f0aa28278652 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:23.903860977Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:5974989cda2655deb3052d601e7406eaf3caee0e7999a6eff251760aac19c199 not found" id=4b0a74f3-6a59-457b-9fa1-f0aa28278652 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:23.904679733Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:5974989cda2655deb3052d601e7406eaf3caee0e7999a6eff251760aac19c199" id=e6b480ae-b753-4e7d-ac82-b2334fb7b887 name=/runtime.v1.ImageService/PullImage Jan 05 07:29:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:23.906886498Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:5974989cda2655deb3052d601e7406eaf3caee0e7999a6eff251760aac19c199\"" Jan 05 07:29:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:24.058419 1425 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-dns/dns-default-2pfzf] Jan 05 07:29:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: 2023-01-05T07:29:24Z [verbose] Add: openshift-dns:dns-default-2pfzf:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/7b5530ed-9b90-41d9-b3c4-0a53de335ef7"}],"ips":[{"version":"4","interface":0,"address":"10.129.2.2/23","gateway":"10.129.2.1"}],"dns":{}} Jan 05 07:29:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: I0105 07:29:24.033028 3125 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-dns", Name:"dns-default-2pfzf", UID:"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63", APIVersion:"v1", ResourceVersion:"35729", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.129.2.2/23] from openshift-sdn Jan 05 07:29:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:24.061418670Z" level=info msg="Got pod network &{Name:dns-default-2pfzf Namespace:openshift-dns ID:81b981c00eed3ce8e7b961151a721e4be9ef2eef83f4433fad233b19e63776bf UID:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 NetNS:/var/run/netns/7b5530ed-9b90-41d9-b3c4-0a53de335ef7 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 07:29:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:24.061591854Z" level=info msg="Checking pod openshift-dns_dns-default-2pfzf for CNI network multus-cni-network (type=multus)" Jan 05 07:29:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 07:29:24.069822 1425 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1cc2aadb_14f2_4d6e_97fc_6fdab8889a63.slice/crio-81b981c00eed3ce8e7b961151a721e4be9ef2eef83f4433fad233b19e63776bf.scope WatchSource:0}: Error finding container 81b981c00eed3ce8e7b961151a721e4be9ef2eef83f4433fad233b19e63776bf: Status 404 returned error can't find the container with id 81b981c00eed3ce8e7b961151a721e4be9ef2eef83f4433fad233b19e63776bf Jan 05 07:29:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:24.074217672Z" level=info msg="Ran pod sandbox 81b981c00eed3ce8e7b961151a721e4be9ef2eef83f4433fad233b19e63776bf with infra container: openshift-dns/dns-default-2pfzf/POD" id=00814159-282e-4ca6-86b0-cb7cabd608b3 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:29:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:24.075628273Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f674b597694b13b7060c9ee611a403f1f87a98e05682c33f9a86140061ee9d9e" id=bf77c9e7-3041-4249-aae2-a5038f85a761 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:24.075835531Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f674b597694b13b7060c9ee611a403f1f87a98e05682c33f9a86140061ee9d9e not found" id=bf77c9e7-3041-4249-aae2-a5038f85a761 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:24.076451943Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f674b597694b13b7060c9ee611a403f1f87a98e05682c33f9a86140061ee9d9e" id=e1d25592-fafa-4854-a09f-a4e40516ce93 name=/runtime.v1.ImageService/PullImage Jan 05 07:29:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:24.077340652Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f674b597694b13b7060c9ee611a403f1f87a98e05682c33f9a86140061ee9d9e\"" Jan 05 07:29:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:24.166582924Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f674b597694b13b7060c9ee611a403f1f87a98e05682c33f9a86140061ee9d9e\"" Jan 05 07:29:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:24.521957 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-8xrbh" event=&{ID:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748 Type:ContainerStarted Data:3e1262384ce877333f952ac7ba12d68b4133181f637b84095f42300e1641af9e} Jan 05 07:29:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:24.522847 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-2pfzf" event=&{ID:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 Type:ContainerStarted Data:81b981c00eed3ce8e7b961151a721e4be9ef2eef83f4433fad233b19e63776bf} Jan 05 07:29:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:25.003936344Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:5974989cda2655deb3052d601e7406eaf3caee0e7999a6eff251760aac19c199\"" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.478206092Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d5fc161170fcf68900b2694039803a2e7f1b31eb913566139a41f91851f166d7" id=ce0638ec-3b72-4541-8b73-45810c7d764c name=/runtime.v1.ImageService/PullImage Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.479671897Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d5fc161170fcf68900b2694039803a2e7f1b31eb913566139a41f91851f166d7" id=b2213d5a-d487-446c-8728-cc2a7252a353 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.481462809Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:852cc803756e640141d6c3eb3e6607f2b0f7076ea77f9190b1f490288f40b3d8,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d5fc161170fcf68900b2694039803a2e7f1b31eb913566139a41f91851f166d7],Size_:574516705,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b2213d5a-d487-446c-8728-cc2a7252a353 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.483036132Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/cni-plugins" id=39795a60-7bcd-49cb-bd9a-8f41f6fda637 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.483144007Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.696856963Z" level=info msg="Created container 3796b8fb7fe0ea062504873d7628979f31aa6c7fb9c113aea6eb192721b2f33d: openshift-multus/multus-additional-cni-plugins-5fgxc/cni-plugins" id=39795a60-7bcd-49cb-bd9a-8f41f6fda637 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.697394717Z" level=info msg="Starting container: 3796b8fb7fe0ea062504873d7628979f31aa6c7fb9c113aea6eb192721b2f33d" id=210df865-8336-456e-bdbc-93afc07f9c2a name=/runtime.v1.RuntimeService/StartContainer Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.705838930Z" level=info msg="Started container" PID=3487 containerID=3796b8fb7fe0ea062504873d7628979f31aa6c7fb9c113aea6eb192721b2f33d description=openshift-multus/multus-additional-cni-plugins-5fgxc/cni-plugins id=210df865-8336-456e-bdbc-93afc07f9c2a name=/runtime.v1.RuntimeService/StartContainer sandboxID=0b984f430b21595a59efe2294b0b9fce6559367942c6743e20dbab9826dae909 Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.718099291Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_345ec096-405b-4ccd-927e-fc23a48d1269\"" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.757042776Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.757080196Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.783716146Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/bandwidth\"" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.796576950Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.796604769Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.796620475Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/bridge\"" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.818486302Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.818519894Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.818536465Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/dhcp\"" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.843470564Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.843496659Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.843513573Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/firewall\"" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.858235727Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.858261343Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.858285543Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/host-device\"" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.872051936Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.872090541Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.872126407Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/host-local\"" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.886000185Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.886020280Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.886048896Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/ipvlan\"" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.897897965Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.897919093Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.897934244Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/loopback\"" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.906136373Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.906154889Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.906168060Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/macvlan\"" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.914592212Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.914611022Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.914624445Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/portmap\"" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.930072340Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.930099348Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.930115213Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/ptp\"" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.940730276Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.940765927Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.940779705Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/sbr\"" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.953089547Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.953108144Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.953122100Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/static\"" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.962937801Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.962956246Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.962970024Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/tuning\"" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.971899442Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.971918050Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.971931715Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/vlan\"" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.982263262Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.982281056Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.982294064Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/vrf\"" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.991877112Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.991895079Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 07:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:40.991909588Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_345ec096-405b-4ccd-927e-fc23a48d1269\"" Jan 05 07:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:41.587712 1425 generic.go:296] "Generic (PLEG): container finished" podID=c0e47249-a5ff-460e-958d-53577387c154 containerID="3796b8fb7fe0ea062504873d7628979f31aa6c7fb9c113aea6eb192721b2f33d" exitCode=0 Jan 05 07:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:41.587747 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerDied Data:3796b8fb7fe0ea062504873d7628979f31aa6c7fb9c113aea6eb192721b2f33d} Jan 05 07:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:41.588495498Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7ad3b9824ea195f38b6346262f0c609415c487ffaeb554725621a8327266da0f" id=2880eed8-5386-44a3-9f0e-b96f25494f62 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:41.588751243Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7ad3b9824ea195f38b6346262f0c609415c487ffaeb554725621a8327266da0f not found" id=2880eed8-5386-44a3-9f0e-b96f25494f62 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:41.589800263Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7ad3b9824ea195f38b6346262f0c609415c487ffaeb554725621a8327266da0f" id=36b4cb93-b2ef-417a-8020-93d0069a0631 name=/runtime.v1.ImageService/PullImage Jan 05 07:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:41.594824150Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7ad3b9824ea195f38b6346262f0c609415c487ffaeb554725621a8327266da0f\"" Jan 05 07:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:41.672341458Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f674b597694b13b7060c9ee611a403f1f87a98e05682c33f9a86140061ee9d9e" id=e1d25592-fafa-4854-a09f-a4e40516ce93 name=/runtime.v1.ImageService/PullImage Jan 05 07:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:41.673237226Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f674b597694b13b7060c9ee611a403f1f87a98e05682c33f9a86140061ee9d9e" id=bae70bef-d29f-4257-937d-e6b126258692 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:41.674823690Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:cdb247afcbd9968f1a59680dcc4de0364f7ddd8fc25dd3a12db59f05c8e026a0,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f674b597694b13b7060c9ee611a403f1f87a98e05682c33f9a86140061ee9d9e],Size_:404212434,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=bae70bef-d29f-4257-937d-e6b126258692 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:41.675717680Z" level=info msg="Creating container: openshift-dns/dns-default-2pfzf/dns" id=8e3c324a-8a97-485e-a04f-e9cb3e8b563b name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:41.675817652Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:41.924196144Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7ad3b9824ea195f38b6346262f0c609415c487ffaeb554725621a8327266da0f\"" Jan 05 07:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:41.959040769Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=ca380ba9-ccf3-42b9-a431-dd14eaa3b18a name=/runtime.v1.ImageService/PullImage Jan 05 07:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:41.959873840Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=e35cd321-2cf8-4956-b11f-8e34a0e20f69 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:41.960064510Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=e35cd321-2cf8-4956-b11f-8e34a0e20f69 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:41.961175105Z" level=info msg="Creating container: openshift-sdn/sdn-fbccx/kube-rbac-proxy" id=325f10a5-680c-49a9-88ac-125de316b70b name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:41.961326754Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:41.988203958Z" level=info msg="Created container e510b9127a187a68744a288407216f7bbd76dfd74cc9d05d826a3e7cc3ecba24: openshift-dns/dns-default-2pfzf/dns" id=8e3c324a-8a97-485e-a04f-e9cb3e8b563b name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:41.988917233Z" level=info msg="Starting container: e510b9127a187a68744a288407216f7bbd76dfd74cc9d05d826a3e7cc3ecba24" id=571660c7-b726-4455-bd61-ab180b72f18e name=/runtime.v1.RuntimeService/StartContainer Jan 05 07:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:42.012065106Z" level=info msg="Started container" PID=3644 containerID=e510b9127a187a68744a288407216f7bbd76dfd74cc9d05d826a3e7cc3ecba24 description=openshift-dns/dns-default-2pfzf/dns id=571660c7-b726-4455-bd61-ab180b72f18e name=/runtime.v1.RuntimeService/StartContainer sandboxID=81b981c00eed3ce8e7b961151a721e4be9ef2eef83f4433fad233b19e63776bf Jan 05 07:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:42.058930169Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=4015f726-6c58-46ac-8dcc-56f70823f5f1 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:42.059226269Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=4015f726-6c58-46ac-8dcc-56f70823f5f1 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:42.064711405Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=2a748232-c094-4845-8fe5-d5f98e3b5b2b name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:42.064916828Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=2a748232-c094-4845-8fe5-d5f98e3b5b2b name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:42.067966881Z" level=info msg="Creating container: openshift-dns/dns-default-2pfzf/kube-rbac-proxy" id=1024a4fd-ce11-47d2-aad8-1fda45241ff8 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:42.068097798Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:42.222598430Z" level=info msg="Created container a1087241f3735617d55881840f0a85549fa47fcf3dacd47d7eb2c41c8d2d4d78: openshift-sdn/sdn-fbccx/kube-rbac-proxy" id=325f10a5-680c-49a9-88ac-125de316b70b name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:42.230327050Z" level=info msg="Starting container: a1087241f3735617d55881840f0a85549fa47fcf3dacd47d7eb2c41c8d2d4d78" id=7acad37d-f233-4871-ac77-5cd3fc378f95 name=/runtime.v1.RuntimeService/StartContainer Jan 05 07:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:42.309148610Z" level=info msg="Started container" PID=3681 containerID=a1087241f3735617d55881840f0a85549fa47fcf3dacd47d7eb2c41c8d2d4d78 description=openshift-sdn/sdn-fbccx/kube-rbac-proxy id=7acad37d-f233-4871-ac77-5cd3fc378f95 name=/runtime.v1.RuntimeService/StartContainer sandboxID=77dc8fbfc80728a344c03166326f04c57592cd6f9e2adc95515af9c2147cc500 Jan 05 07:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:42.425757768Z" level=info msg="Created container 6590094f03e48268180ff9efb6a82858a36859f46f26c427f22134073ffa60ed: openshift-dns/dns-default-2pfzf/kube-rbac-proxy" id=1024a4fd-ce11-47d2-aad8-1fda45241ff8 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:42.426449761Z" level=info msg="Starting container: 6590094f03e48268180ff9efb6a82858a36859f46f26c427f22134073ffa60ed" id=d850eeea-3c5a-401e-89cb-b09a4e890155 name=/runtime.v1.RuntimeService/StartContainer Jan 05 07:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:42.434359301Z" level=info msg="Started container" PID=3723 containerID=6590094f03e48268180ff9efb6a82858a36859f46f26c427f22134073ffa60ed description=openshift-dns/dns-default-2pfzf/kube-rbac-proxy id=d850eeea-3c5a-401e-89cb-b09a4e890155 name=/runtime.v1.RuntimeService/StartContainer sandboxID=81b981c00eed3ce8e7b961151a721e4be9ef2eef83f4433fad233b19e63776bf Jan 05 07:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:42.594021 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-2pfzf" event=&{ID:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 Type:ContainerStarted Data:6590094f03e48268180ff9efb6a82858a36859f46f26c427f22134073ffa60ed} Jan 05 07:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:42.594056 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-2pfzf" event=&{ID:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 Type:ContainerStarted Data:e510b9127a187a68744a288407216f7bbd76dfd74cc9d05d826a3e7cc3ecba24} Jan 05 07:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:42.594834 1425 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-2pfzf" Jan 05 07:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:42.596494 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-sdn/sdn-fbccx" event=&{ID:b5d621a5-07d9-459b-8745-a991cf4a179e Type:ContainerStarted Data:a1087241f3735617d55881840f0a85549fa47fcf3dacd47d7eb2c41c8d2d4d78} Jan 05 07:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:42.596762 1425 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-sdn/sdn-fbccx" Jan 05 07:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:42.709336 1425 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-sdn/sdn-fbccx" Jan 05 07:29:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:46.961702578Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:5974989cda2655deb3052d601e7406eaf3caee0e7999a6eff251760aac19c199" id=e6b480ae-b753-4e7d-ac82-b2334fb7b887 name=/runtime.v1.ImageService/PullImage Jan 05 07:29:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:46.963663131Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:5974989cda2655deb3052d601e7406eaf3caee0e7999a6eff251760aac19c199" id=f678c717-7caf-44b8-8f7b-d6ee749803a1 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:46.965611279Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:c34c97b229e46faa94bab6ad2559fd683bd036df06031c9097dd27f7daaf96c7,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:5974989cda2655deb3052d601e7406eaf3caee0e7999a6eff251760aac19c199],Size_:417780553,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=f678c717-7caf-44b8-8f7b-d6ee749803a1 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:46.967756608Z" level=info msg="Creating container: openshift-ingress-canary/ingress-canary-8xrbh/serve-healthcheck-canary" id=b57381d6-ea10-4348-9fb2-64a7715fbad3 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:29:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:46.967890907Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:46.999140684Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=50bad22e-2c38-408b-b3e8-022678576f7f name=/runtime.v1.ImageService/PullImage Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:47.009024077Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=a581303a-9e66-4c71-9454-27ea50a99302 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:47.010685320Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=a581303a-9e66-4c71-9454-27ea50a99302 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:47.016662803Z" level=info msg="Creating container: openshift-machine-config-operator/machine-config-daemon-hf8f5/oauth-proxy" id=abd41323-c399-42c9-b5ee-a4c1cec8717d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:47.016764780Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:47.251032000Z" level=info msg="Created container 3a859bd8e5377b903866f4458a92d4124139c755f651682a841008143e51b0f5: openshift-ingress-canary/ingress-canary-8xrbh/serve-healthcheck-canary" id=b57381d6-ea10-4348-9fb2-64a7715fbad3 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:47.251868645Z" level=info msg="Starting container: 3a859bd8e5377b903866f4458a92d4124139c755f651682a841008143e51b0f5" id=198348ec-990f-401c-bca7-4fc308c41845 name=/runtime.v1.RuntimeService/StartContainer Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:47.265784660Z" level=info msg="Started container" PID=3911 containerID=3a859bd8e5377b903866f4458a92d4124139c755f651682a841008143e51b0f5 description=openshift-ingress-canary/ingress-canary-8xrbh/serve-healthcheck-canary id=198348ec-990f-401c-bca7-4fc308c41845 name=/runtime.v1.RuntimeService/StartContainer sandboxID=3e1262384ce877333f952ac7ba12d68b4133181f637b84095f42300e1641af9e Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:47.278974465Z" level=info msg="Created container aa9d11d2a5b9257520e59987ba2d4c9dbb807692557e3dc9a20e4b196502fc21: openshift-machine-config-operator/machine-config-daemon-hf8f5/oauth-proxy" id=abd41323-c399-42c9-b5ee-a4c1cec8717d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:47.286629259Z" level=info msg="Starting container: aa9d11d2a5b9257520e59987ba2d4c9dbb807692557e3dc9a20e4b196502fc21" id=9d42b818-b39a-4196-992a-0725ca353e1f name=/runtime.v1.RuntimeService/StartContainer Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:47.314005914Z" level=info msg="Started container" PID=3918 containerID=aa9d11d2a5b9257520e59987ba2d4c9dbb807692557e3dc9a20e4b196502fc21 description=openshift-machine-config-operator/machine-config-daemon-hf8f5/oauth-proxy id=9d42b818-b39a-4196-992a-0725ca353e1f name=/runtime.v1.RuntimeService/StartContainer sandboxID=2cd51e2f6a1cbcdeb26d92c54070ab2a85b957f76038eb175b75e3f2ae639ed7 Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:47.348505234Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7ad3b9824ea195f38b6346262f0c609415c487ffaeb554725621a8327266da0f" id=36b4cb93-b2ef-417a-8020-93d0069a0631 name=/runtime.v1.ImageService/PullImage Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:47.351699948Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7ad3b9824ea195f38b6346262f0c609415c487ffaeb554725621a8327266da0f" id=eeb76236-09cb-4ab2-b98b-e033368b0315 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:47.352996141Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:4c7cce1e79d06b3bc147fc59bd0a72991be97e767193f661db48974f3bb9ac30,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7ad3b9824ea195f38b6346262f0c609415c487ffaeb554725621a8327266da0f],Size_:336119748,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=eeb76236-09cb-4ab2-b98b-e033368b0315 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:47.353785549Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/bond-cni-plugin" id=ecb8d239-99ea-4e6c-a52f-bb2ba6affd87 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:47.353896816Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:47.547019779Z" level=info msg="Created container a188d9a4b8d44a2cee021025b65e2185a5ddc0fe47f56a966e051b665601a335: openshift-multus/multus-additional-cni-plugins-5fgxc/bond-cni-plugin" id=ecb8d239-99ea-4e6c-a52f-bb2ba6affd87 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:47.547667057Z" level=info msg="Starting container: a188d9a4b8d44a2cee021025b65e2185a5ddc0fe47f56a966e051b665601a335" id=162c09a7-bd20-44c5-8806-7f24523195ba name=/runtime.v1.RuntimeService/StartContainer Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:47.555086081Z" level=info msg="Started container" PID=3982 containerID=a188d9a4b8d44a2cee021025b65e2185a5ddc0fe47f56a966e051b665601a335 description=openshift-multus/multus-additional-cni-plugins-5fgxc/bond-cni-plugin id=162c09a7-bd20-44c5-8806-7f24523195ba name=/runtime.v1.RuntimeService/StartContainer sandboxID=0b984f430b21595a59efe2294b0b9fce6559367942c6743e20dbab9826dae909 Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:47.565291004Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_571b1302-e5a3-4363-b6a6-4ceafbfc5683\"" Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:47.578014000Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:47.578040858Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:47.578212925Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/bond\"" Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:47.588905941Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:47.588931590Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:47.588949586Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_571b1302-e5a3-4363-b6a6-4ceafbfc5683\"" Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:47.615296 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-8xrbh" event=&{ID:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748 Type:ContainerStarted Data:3a859bd8e5377b903866f4458a92d4124139c755f651682a841008143e51b0f5} Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:47.617574 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" event=&{ID:997b7ab4-939e-465c-9c7d-4a2ebe3a797a Type:ContainerStarted Data:aa9d11d2a5b9257520e59987ba2d4c9dbb807692557e3dc9a20e4b196502fc21} Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:47.620241 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerStarted Data:a188d9a4b8d44a2cee021025b65e2185a5ddc0fe47f56a966e051b665601a335} Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:47.620905211Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:a8bcc8cf428e222c7e6677fe5f7f05e79da2aed1ded851cad696d83c94860b4c" id=b4905c8a-4a7e-4905-9efc-d8f9b0b86ebb name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:47.621108020Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:a8bcc8cf428e222c7e6677fe5f7f05e79da2aed1ded851cad696d83c94860b4c not found" id=b4905c8a-4a7e-4905-9efc-d8f9b0b86ebb name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:47.621857310Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:a8bcc8cf428e222c7e6677fe5f7f05e79da2aed1ded851cad696d83c94860b4c" id=30da4902-2aaa-414b-826e-f3b24b24ba63 name=/runtime.v1.ImageService/PullImage Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:47.622757570Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:a8bcc8cf428e222c7e6677fe5f7f05e79da2aed1ded851cad696d83c94860b4c\"" Jan 05 07:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:47.707822007Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:a8bcc8cf428e222c7e6677fe5f7f05e79da2aed1ded851cad696d83c94860b4c\"" Jan 05 07:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:48.623144 1425 generic.go:296] "Generic (PLEG): container finished" podID=c0e47249-a5ff-460e-958d-53577387c154 containerID="a188d9a4b8d44a2cee021025b65e2185a5ddc0fe47f56a966e051b665601a335" exitCode=0 Jan 05 07:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:48.623910 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerDied Data:a188d9a4b8d44a2cee021025b65e2185a5ddc0fe47f56a966e051b665601a335} Jan 05 07:29:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:54.809729 1425 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-2pfzf" Jan 05 07:29:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:58.577994220Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:a8bcc8cf428e222c7e6677fe5f7f05e79da2aed1ded851cad696d83c94860b4c" id=30da4902-2aaa-414b-826e-f3b24b24ba63 name=/runtime.v1.ImageService/PullImage Jan 05 07:29:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:58.578846238Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:a8bcc8cf428e222c7e6677fe5f7f05e79da2aed1ded851cad696d83c94860b4c" id=87986ce4-fa3b-4aa8-860e-248531dadbe9 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:58.580192433Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:34fa51fa29c1ebd2a3ad620255bb70edf89b0a96348f08efac3dc2ca3200602d,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:a8bcc8cf428e222c7e6677fe5f7f05e79da2aed1ded851cad696d83c94860b4c],Size_:275275974,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=87986ce4-fa3b-4aa8-860e-248531dadbe9 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:58.580968649Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/routeoverride-cni" id=a8e92d88-1408-4705-90f5-f1ad7669285b name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:29:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:58.581062420Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:29:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:58.742281313Z" level=info msg="Created container 04a6f9fdd6aa35a6404a14efcf5654dfe48f0e4aa4f114bb652ef14f217c6cc7: openshift-multus/multus-additional-cni-plugins-5fgxc/routeoverride-cni" id=a8e92d88-1408-4705-90f5-f1ad7669285b name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:29:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:58.742730032Z" level=info msg="Starting container: 04a6f9fdd6aa35a6404a14efcf5654dfe48f0e4aa4f114bb652ef14f217c6cc7" id=95c9d438-347f-4756-8f78-ff5e03c094fd name=/runtime.v1.RuntimeService/StartContainer Jan 05 07:29:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:58.749478249Z" level=info msg="Started container" PID=4295 containerID=04a6f9fdd6aa35a6404a14efcf5654dfe48f0e4aa4f114bb652ef14f217c6cc7 description=openshift-multus/multus-additional-cni-plugins-5fgxc/routeoverride-cni id=95c9d438-347f-4756-8f78-ff5e03c094fd name=/runtime.v1.RuntimeService/StartContainer sandboxID=0b984f430b21595a59efe2294b0b9fce6559367942c6743e20dbab9826dae909 Jan 05 07:29:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:58.764381691Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_9d875156-2b3f-4305-9710-9fa7a8ae2832\"" Jan 05 07:29:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:58.778079189Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 07:29:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:58.778102904Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 07:29:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:58.778120305Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/route-override\"" Jan 05 07:29:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:58.787623848Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 07:29:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:58.787648223Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 07:29:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:58.789003133Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_9d875156-2b3f-4305-9710-9fa7a8ae2832\"" Jan 05 07:29:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:59.662563 1425 generic.go:296] "Generic (PLEG): container finished" podID=c0e47249-a5ff-460e-958d-53577387c154 containerID="04a6f9fdd6aa35a6404a14efcf5654dfe48f0e4aa4f114bb652ef14f217c6cc7" exitCode=0 Jan 05 07:29:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:29:59.662602 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerDied Data:04a6f9fdd6aa35a6404a14efcf5654dfe48f0e4aa4f114bb652ef14f217c6cc7} Jan 05 07:29:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:59.663514852Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451" id=2352a2b8-4fc1-4bc1-b0cd-df6610ec0bc9 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:59.663766951Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451 not found" id=2352a2b8-4fc1-4bc1-b0cd-df6610ec0bc9 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:29:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:59.664467494Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451" id=9221b75e-85ef-4025-83c9-af0f0bfeb8c9 name=/runtime.v1.ImageService/PullImage Jan 05 07:29:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:59.665342613Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451\"" Jan 05 07:29:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:29:59.762735154Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451\"" Jan 05 07:30:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:30:04.445756 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-mmxtc\" (UniqueName: \"kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc\") pod \"network-check-target-xmq2g\" (UID: \"f364a949-09d7-4cfb-83ff-e532b822a557\") " pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 07:30:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:30:04.445810 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 07:30:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:30:04.459422 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 07:30:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:30:04.475268 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmxtc\" (UniqueName: \"kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc\") pod \"network-check-target-xmq2g\" (UID: \"f364a949-09d7-4cfb-83ff-e532b822a557\") " pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 07:30:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:30:04.608583 1425 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 07:30:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:04.609043916Z" level=info msg="Running pod sandbox: openshift-multus/network-metrics-daemon-hq594/POD" id=d7635cbb-5f5a-46f6-8d92-a1a6a1d45a1f name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:30:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:04.609098025Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:30:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:04.627999842Z" level=info msg="Got pod network &{Name:network-metrics-daemon-hq594 Namespace:openshift-multus ID:5fc6d6ddb9ba864a50f6934237ca81ca5c56501c56049bb7190d5ac76187ea49 UID:50152296-9489-4fa8-aa42-7858debe1c08 NetNS:/var/run/netns/444697ac-5441-4dc1-82e2-47eb57447726 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 07:30:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:04.628039213Z" level=info msg="Adding pod openshift-multus_network-metrics-daemon-hq594 to CNI network \"multus-cni-network\" (type=multus)" Jan 05 07:30:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:30:04.698275 1425 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 07:30:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:04.702733176Z" level=info msg="Running pod sandbox: openshift-network-diagnostics/network-check-target-xmq2g/POD" id=05057488-9589-4b69-bd7b-ba1126408fed name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:30:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:04.702800416Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:30:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:04.742411675Z" level=info msg="Got pod network &{Name:network-check-target-xmq2g Namespace:openshift-network-diagnostics ID:e1cf6641b5c3c13478f683a4318f93d2bf7b7d9376939506acf8a588476f10eb UID:f364a949-09d7-4cfb-83ff-e532b822a557 NetNS:/var/run/netns/ecce2622-252b-4599-9586-1dab87f71a92 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 07:30:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:04.742451147Z" level=info msg="Adding pod openshift-network-diagnostics_network-check-target-xmq2g to CNI network \"multus-cni-network\" (type=multus)" Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:30:06.115716 1425 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-network-diagnostics/network-check-target-xmq2g] Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: 2023-01-05T07:30:06Z [verbose] Add: openshift-network-diagnostics:network-check-target-xmq2g:f364a949-09d7-4cfb-83ff-e532b822a557:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/ecce2622-252b-4599-9586-1dab87f71a92"}],"ips":[{"version":"4","interface":0,"address":"10.129.2.5/23","gateway":"10.129.2.1"}],"dns":{}} Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: I0105 07:30:06.094570 4440 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-network-diagnostics", Name:"network-check-target-xmq2g", UID:"f364a949-09d7-4cfb-83ff-e532b822a557", APIVersion:"v1", ResourceVersion:"34852", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.129.2.5/23] from openshift-sdn Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:06.118836483Z" level=info msg="Got pod network &{Name:network-check-target-xmq2g Namespace:openshift-network-diagnostics ID:e1cf6641b5c3c13478f683a4318f93d2bf7b7d9376939506acf8a588476f10eb UID:f364a949-09d7-4cfb-83ff-e532b822a557 NetNS:/var/run/netns/ecce2622-252b-4599-9586-1dab87f71a92 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:06.118972323Z" level=info msg="Checking pod openshift-network-diagnostics_network-check-target-xmq2g for CNI network multus-cni-network (type=multus)" Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: 2023-01-05T07:30:06Z [verbose] Add: openshift-multus:network-metrics-daemon-hq594:50152296-9489-4fa8-aa42-7858debe1c08:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/444697ac-5441-4dc1-82e2-47eb57447726"}],"ips":[{"version":"4","interface":0,"address":"10.129.2.4/23","gateway":"10.129.2.1"}],"dns":{}} Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: I0105 07:30:06.173629 4414 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-multus", Name:"network-metrics-daemon-hq594", UID:"50152296-9489-4fa8-aa42-7858debe1c08", APIVersion:"v1", ResourceVersion:"34866", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.129.2.4/23] from openshift-sdn Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:06.197774124Z" level=info msg="Got pod network &{Name:network-metrics-daemon-hq594 Namespace:openshift-multus ID:5fc6d6ddb9ba864a50f6934237ca81ca5c56501c56049bb7190d5ac76187ea49 UID:50152296-9489-4fa8-aa42-7858debe1c08 NetNS:/var/run/netns/444697ac-5441-4dc1-82e2-47eb57447726 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:06.197925011Z" level=info msg="Checking pod openshift-multus_network-metrics-daemon-hq594 for CNI network multus-cni-network (type=multus)" Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:30:06.199606 1425 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-multus/network-metrics-daemon-hq594] Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 07:30:06.400343 1425 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf364a949_09d7_4cfb_83ff_e532b822a557.slice/crio-e1cf6641b5c3c13478f683a4318f93d2bf7b7d9376939506acf8a588476f10eb.scope WatchSource:0}: Error finding container e1cf6641b5c3c13478f683a4318f93d2bf7b7d9376939506acf8a588476f10eb: Status 404 returned error can't find the container with id e1cf6641b5c3c13478f683a4318f93d2bf7b7d9376939506acf8a588476f10eb Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:06.402034846Z" level=info msg="Ran pod sandbox 5fc6d6ddb9ba864a50f6934237ca81ca5c56501c56049bb7190d5ac76187ea49 with infra container: openshift-multus/network-metrics-daemon-hq594/POD" id=d7635cbb-5f5a-46f6-8d92-a1a6a1d45a1f name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:06.405114891Z" level=info msg="Ran pod sandbox e1cf6641b5c3c13478f683a4318f93d2bf7b7d9376939506acf8a588476f10eb with infra container: openshift-network-diagnostics/network-check-target-xmq2g/POD" id=05057488-9589-4b69-bd7b-ba1126408fed name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:06.406696460Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9e33552d1aaea9cf6a301fa438cfc8305826c5ddb58ac8437d884e55e71c6e44" id=15a93abe-0ae0-4a1d-b616-549d5e8b31c6 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:06.407150679Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7" id=1227085f-26ab-409f-8fcd-6754a7593da8 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:06.417826429Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9e33552d1aaea9cf6a301fa438cfc8305826c5ddb58ac8437d884e55e71c6e44 not found" id=15a93abe-0ae0-4a1d-b616-549d5e8b31c6 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:06.418495616Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9e33552d1aaea9cf6a301fa438cfc8305826c5ddb58ac8437d884e55e71c6e44" id=fc6f6011-e884-4163-a7fa-44ee8ad05a3b name=/runtime.v1.ImageService/PullImage Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:06.421567469Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9e33552d1aaea9cf6a301fa438cfc8305826c5ddb58ac8437d884e55e71c6e44\"" Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:06.422691787Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7 not found" id=1227085f-26ab-409f-8fcd-6754a7593da8 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:06.423352485Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7" id=863af58c-9081-410f-ba4a-6c33f56e5d1c name=/runtime.v1.ImageService/PullImage Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:06.424667719Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7\"" Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:06.435803900Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451" id=9221b75e-85ef-4025-83c9-af0f0bfeb8c9 name=/runtime.v1.ImageService/PullImage Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:06.436418733Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451" id=43788c9d-27d9-4c98-a4b0-9beeb41c8cae name=/runtime.v1.ImageService/ImageStatus Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:06.437978253Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:56f4df3c126a569448a12d5112b7c20e99e219b2b2fe6a04c4ba743396fd7751,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451],Size_:445620510,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=43788c9d-27d9-4c98-a4b0-9beeb41c8cae name=/runtime.v1.ImageService/ImageStatus Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:06.438886425Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/whereabouts-cni-bincopy" id=a27c4239-5b6b-40ef-9bb4-759366f5f6d8 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:06.438975490Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:06.531176075Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7\"" Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:06.555680722Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9e33552d1aaea9cf6a301fa438cfc8305826c5ddb58ac8437d884e55e71c6e44\"" Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:06.613045172Z" level=info msg="Created container dcbb43ac5bfa337da65ed281387b738e4c0e97beb89b5fbcfd124d7c959f534b: openshift-multus/multus-additional-cni-plugins-5fgxc/whereabouts-cni-bincopy" id=a27c4239-5b6b-40ef-9bb4-759366f5f6d8 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:06.613614528Z" level=info msg="Starting container: dcbb43ac5bfa337da65ed281387b738e4c0e97beb89b5fbcfd124d7c959f534b" id=7a6e734e-5215-4589-ba50-ddc90fd736d2 name=/runtime.v1.RuntimeService/StartContainer Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:06.622604280Z" level=info msg="Started container" PID=4524 containerID=dcbb43ac5bfa337da65ed281387b738e4c0e97beb89b5fbcfd124d7c959f534b description=openshift-multus/multus-additional-cni-plugins-5fgxc/whereabouts-cni-bincopy id=7a6e734e-5215-4589-ba50-ddc90fd736d2 name=/runtime.v1.RuntimeService/StartContainer sandboxID=0b984f430b21595a59efe2294b0b9fce6559367942c6743e20dbab9826dae909 Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:06.630628126Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_50bbbd64-7eb6-4e55-8295-fa05c17c2078\"" Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:06.667032464Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:06.667078240Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:06.667869801Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/whereabouts\"" Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:06.682043186Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:06.682069620Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:06.683625594Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_50bbbd64-7eb6-4e55-8295-fa05c17c2078\"" Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:30:06.684419 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xmq2g" event=&{ID:f364a949-09d7-4cfb-83ff-e532b822a557 Type:ContainerStarted Data:e1cf6641b5c3c13478f683a4318f93d2bf7b7d9376939506acf8a588476f10eb} Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:30:06.685954 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hq594" event=&{ID:50152296-9489-4fa8-aa42-7858debe1c08 Type:ContainerStarted Data:5fc6d6ddb9ba864a50f6934237ca81ca5c56501c56049bb7190d5ac76187ea49} Jan 05 07:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:30:06.704065 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerStarted Data:dcbb43ac5bfa337da65ed281387b738e4c0e97beb89b5fbcfd124d7c959f534b} Jan 05 07:30:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:30:07.716363 1425 generic.go:296] "Generic (PLEG): container finished" podID=c0e47249-a5ff-460e-958d-53577387c154 containerID="dcbb43ac5bfa337da65ed281387b738e4c0e97beb89b5fbcfd124d7c959f534b" exitCode=0 Jan 05 07:30:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:30:07.716403 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerDied Data:dcbb43ac5bfa337da65ed281387b738e4c0e97beb89b5fbcfd124d7c959f534b} Jan 05 07:30:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:07.717074357Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451" id=ee860cb7-d3ae-4f59-a633-abe7b52a35a2 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:30:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:07.719690796Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:56f4df3c126a569448a12d5112b7c20e99e219b2b2fe6a04c4ba743396fd7751,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451],Size_:445620510,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=ee860cb7-d3ae-4f59-a633-abe7b52a35a2 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:30:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:07.720667600Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451" id=aa05bdec-bc3a-4e46-99b8-27abd5e95af2 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:30:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:07.722818483Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:56f4df3c126a569448a12d5112b7c20e99e219b2b2fe6a04c4ba743396fd7751,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451],Size_:445620510,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=aa05bdec-bc3a-4e46-99b8-27abd5e95af2 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:30:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:07.727270873Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/whereabouts-cni" id=197a9887-068a-4ba4-88c7-9aa7955e51cd name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:30:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:07.727399233Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:30:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:07.940576206Z" level=info msg="Created container 915fce4e262fb0e9a416f59233c09e6685e55095a8fcc31024c67f9ad8a94077: openshift-multus/multus-additional-cni-plugins-5fgxc/whereabouts-cni" id=197a9887-068a-4ba4-88c7-9aa7955e51cd name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:30:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:07.941295609Z" level=info msg="Starting container: 915fce4e262fb0e9a416f59233c09e6685e55095a8fcc31024c67f9ad8a94077" id=e3271ccc-4d82-4684-bd4c-9915422cc608 name=/runtime.v1.RuntimeService/StartContainer Jan 05 07:30:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:07.951214061Z" level=info msg="Started container" PID=4622 containerID=915fce4e262fb0e9a416f59233c09e6685e55095a8fcc31024c67f9ad8a94077 description=openshift-multus/multus-additional-cni-plugins-5fgxc/whereabouts-cni id=e3271ccc-4d82-4684-bd4c-9915422cc608 name=/runtime.v1.RuntimeService/StartContainer sandboxID=0b984f430b21595a59efe2294b0b9fce6559367942c6743e20dbab9826dae909 Jan 05 07:30:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:07.952461772Z" level=info msg="CNI monitoring event CREATE \"/etc/kubernetes/cni/net.d/whereabouts.d\"" Jan 05 07:30:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:07.975158678Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 07:30:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:07.975208833Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 07:30:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:30:08.724994 1425 generic.go:296] "Generic (PLEG): container finished" podID=c0e47249-a5ff-460e-958d-53577387c154 containerID="915fce4e262fb0e9a416f59233c09e6685e55095a8fcc31024c67f9ad8a94077" exitCode=0 Jan 05 07:30:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:30:08.725053 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerDied Data:915fce4e262fb0e9a416f59233c09e6685e55095a8fcc31024c67f9ad8a94077} Jan 05 07:30:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:08.726526204Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=f1b734d2-b887-4c22-8d3e-002919850305 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:30:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:08.726732926Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=f1b734d2-b887-4c22-8d3e-002919850305 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:30:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:08.731279813Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=9f696bfc-4a6e-4faa-80a7-179882c30877 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:30:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:08.731474005Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=9f696bfc-4a6e-4faa-80a7-179882c30877 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:30:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:08.732374491Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/kube-multus-additional-cni-plugins" id=11a038cd-89e5-4446-bcda-5b40082934e9 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:30:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:08.732505154Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:30:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:08.831056423Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9e33552d1aaea9cf6a301fa438cfc8305826c5ddb58ac8437d884e55e71c6e44" id=fc6f6011-e884-4163-a7fa-44ee8ad05a3b name=/runtime.v1.ImageService/PullImage Jan 05 07:30:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:08.835690637Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9e33552d1aaea9cf6a301fa438cfc8305826c5ddb58ac8437d884e55e71c6e44" id=5c30d9e6-b14e-49f3-a379-ad9f6916f1f7 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:30:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:08.837921527Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:7b87fc786118817b8913f28a61e01db37b08380f754bbaacf425508b208632a0,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9e33552d1aaea9cf6a301fa438cfc8305826c5ddb58ac8437d884e55e71c6e44],Size_:312894682,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=5c30d9e6-b14e-49f3-a379-ad9f6916f1f7 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:30:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:08.839066073Z" level=info msg="Creating container: openshift-multus/network-metrics-daemon-hq594/network-metrics-daemon" id=275657a4-5262-4241-950d-801385e9ca33 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:30:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:08.839178990Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:30:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:09.022389649Z" level=info msg="Created container eaf52a652c4aa6c5ee402bfb178f1e48fd70ac9ee8b3ef472f9c202e015d1319: openshift-multus/multus-additional-cni-plugins-5fgxc/kube-multus-additional-cni-plugins" id=11a038cd-89e5-4446-bcda-5b40082934e9 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:30:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:09.023376868Z" level=info msg="Starting container: eaf52a652c4aa6c5ee402bfb178f1e48fd70ac9ee8b3ef472f9c202e015d1319" id=acc5d9cc-cee1-469a-885f-e7dde3a4002c name=/runtime.v1.RuntimeService/StartContainer Jan 05 07:30:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:09.035083544Z" level=info msg="Started container" PID=4695 containerID=eaf52a652c4aa6c5ee402bfb178f1e48fd70ac9ee8b3ef472f9c202e015d1319 description=openshift-multus/multus-additional-cni-plugins-5fgxc/kube-multus-additional-cni-plugins id=acc5d9cc-cee1-469a-885f-e7dde3a4002c name=/runtime.v1.RuntimeService/StartContainer sandboxID=0b984f430b21595a59efe2294b0b9fce6559367942c6743e20dbab9826dae909 Jan 05 07:30:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:09.095738382Z" level=info msg="Created container b3b076d3d3254c5a160edc4256aba143f0d09bfc89022f3c8a305f947bffab97: openshift-multus/network-metrics-daemon-hq594/network-metrics-daemon" id=275657a4-5262-4241-950d-801385e9ca33 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:30:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:09.096211798Z" level=info msg="Starting container: b3b076d3d3254c5a160edc4256aba143f0d09bfc89022f3c8a305f947bffab97" id=f52fbe2e-2f90-4370-9ad7-8c4d0784581f name=/runtime.v1.RuntimeService/StartContainer Jan 05 07:30:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:09.104396712Z" level=info msg="Started container" PID=4729 containerID=b3b076d3d3254c5a160edc4256aba143f0d09bfc89022f3c8a305f947bffab97 description=openshift-multus/network-metrics-daemon-hq594/network-metrics-daemon id=f52fbe2e-2f90-4370-9ad7-8c4d0784581f name=/runtime.v1.RuntimeService/StartContainer sandboxID=5fc6d6ddb9ba864a50f6934237ca81ca5c56501c56049bb7190d5ac76187ea49 Jan 05 07:30:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:09.114668420Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=aa4743d0-8416-4fb2-9f2a-3aff65fc7a82 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:30:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:09.114881830Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=aa4743d0-8416-4fb2-9f2a-3aff65fc7a82 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:30:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:09.120097728Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=63422718-6493-4262-b7da-e8c842a9c2bb name=/runtime.v1.ImageService/ImageStatus Jan 05 07:30:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:09.120326075Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=63422718-6493-4262-b7da-e8c842a9c2bb name=/runtime.v1.ImageService/ImageStatus Jan 05 07:30:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:09.121699435Z" level=info msg="Creating container: openshift-multus/network-metrics-daemon-hq594/kube-rbac-proxy" id=f7812512-7988-48a6-90bd-676917b73aa0 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:30:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:09.121823755Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:30:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:09.322202891Z" level=info msg="Created container 034d9fb0f294476dec493bd498c777dbaaa492fa0a5b473a0012f3a04ca661da: openshift-multus/network-metrics-daemon-hq594/kube-rbac-proxy" id=f7812512-7988-48a6-90bd-676917b73aa0 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:30:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:09.326677253Z" level=info msg="Starting container: 034d9fb0f294476dec493bd498c777dbaaa492fa0a5b473a0012f3a04ca661da" id=242699b9-7a20-4f9b-8781-405e2428deed name=/runtime.v1.RuntimeService/StartContainer Jan 05 07:30:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:09.344523631Z" level=info msg="Started container" PID=4769 containerID=034d9fb0f294476dec493bd498c777dbaaa492fa0a5b473a0012f3a04ca661da description=openshift-multus/network-metrics-daemon-hq594/kube-rbac-proxy id=242699b9-7a20-4f9b-8781-405e2428deed name=/runtime.v1.RuntimeService/StartContainer sandboxID=5fc6d6ddb9ba864a50f6934237ca81ca5c56501c56049bb7190d5ac76187ea49 Jan 05 07:30:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:30:09.730147 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerStarted Data:eaf52a652c4aa6c5ee402bfb178f1e48fd70ac9ee8b3ef472f9c202e015d1319} Jan 05 07:30:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:30:09.732342 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hq594" event=&{ID:50152296-9489-4fa8-aa42-7858debe1c08 Type:ContainerStarted Data:034d9fb0f294476dec493bd498c777dbaaa492fa0a5b473a0012f3a04ca661da} Jan 05 07:30:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:30:09.732369 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hq594" event=&{ID:50152296-9489-4fa8-aa42-7858debe1c08 Type:ContainerStarted Data:b3b076d3d3254c5a160edc4256aba143f0d09bfc89022f3c8a305f947bffab97} Jan 05 07:30:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:16.194778298Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7" id=863af58c-9081-410f-ba4a-6c33f56e5d1c name=/runtime.v1.ImageService/PullImage Jan 05 07:30:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:16.195790970Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7" id=d698b9dc-1fff-429d-bb51-2dde3dc7b599 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:30:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:16.197280958Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:4cb41c73a232ec4401c5db32cdb523810dc12ac605cdc6a9ce388e90930eb6c4,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7],Size_:515955107,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=d698b9dc-1fff-429d-bb51-2dde3dc7b599 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:30:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:16.197980867Z" level=info msg="Creating container: openshift-network-diagnostics/network-check-target-xmq2g/network-check-target-container" id=d86b2403-d1e9-465a-a605-f7a856237420 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:30:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:16.198088391Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 07:30:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:16.393102946Z" level=info msg="Created container cff7ae4e0a1dfd7a6a0fa04b6a44c76ca646353a6dd5a14cacd83fa933e46a8f: openshift-network-diagnostics/network-check-target-xmq2g/network-check-target-container" id=d86b2403-d1e9-465a-a605-f7a856237420 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 07:30:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:16.393606140Z" level=info msg="Starting container: cff7ae4e0a1dfd7a6a0fa04b6a44c76ca646353a6dd5a14cacd83fa933e46a8f" id=78c395a2-7bf6-4450-a43a-6c6140437898 name=/runtime.v1.RuntimeService/StartContainer Jan 05 07:30:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:30:16.407097245Z" level=info msg="Started container" PID=4888 containerID=cff7ae4e0a1dfd7a6a0fa04b6a44c76ca646353a6dd5a14cacd83fa933e46a8f description=openshift-network-diagnostics/network-check-target-xmq2g/network-check-target-container id=78c395a2-7bf6-4450-a43a-6c6140437898 name=/runtime.v1.RuntimeService/StartContainer sandboxID=e1cf6641b5c3c13478f683a4318f93d2bf7b7d9376939506acf8a588476f10eb Jan 05 07:30:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:30:16.751627 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xmq2g" event=&{ID:f364a949-09d7-4cfb-83ff-e532b822a557 Type:ContainerStarted Data:cff7ae4e0a1dfd7a6a0fa04b6a44c76ca646353a6dd5a14cacd83fa933e46a8f} Jan 05 07:30:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:30:16.752134 1425 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 07:30:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:30:54.702046 1425 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 07:32:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 07:32:36.368690 1425 kubelet.go:1343] "Image garbage collection succeeded" Jan 05 07:32:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:32:36.422287815Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=e2eed4d3-beec-4466-8fa1-70608c3461a2 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:32:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:32:36.422538223Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=e2eed4d3-beec-4466-8fa1-70608c3461a2 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:37:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:37:36.424711180Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=f2a647df-fa59-4a81-9037-dc5e7432f3c3 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:37:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:37:36.424947429Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=f2a647df-fa59-4a81-9037-dc5e7432f3c3 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:42:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:42:36.427915270Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=d83f29bf-da64-46c7-8045-2464c740c9fc name=/runtime.v1.ImageService/ImageStatus Jan 05 07:42:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:42:36.428195145Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=d83f29bf-da64-46c7-8045-2464c740c9fc name=/runtime.v1.ImageService/ImageStatus Jan 05 07:47:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:47:36.431200947Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=2c1a616c-4164-4e0b-8e41-be9539808360 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:47:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:47:36.431445670Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=2c1a616c-4164-4e0b-8e41-be9539808360 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:52:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:52:36.435125061Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=475c5de5-4d39-44e1-a6d9-c467c3de85e9 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:52:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:52:36.435403191Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=475c5de5-4d39-44e1-a6d9-c467c3de85e9 name=/runtime.v1.ImageService/ImageStatus Jan 05 07:57:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:57:36.437955538Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=d90456ea-d500-4415-a173-11b405d0396c name=/runtime.v1.ImageService/ImageStatus Jan 05 07:57:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 07:57:36.438203429Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=d90456ea-d500-4415-a173-11b405d0396c name=/runtime.v1.ImageService/ImageStatus Jan 05 08:02:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 08:02:36.441346926Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=5c71a1ea-fc0a-43d1-9688-ff1bab454ba7 name=/runtime.v1.ImageService/ImageStatus Jan 05 08:02:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 08:02:36.441623070Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=5c71a1ea-fc0a-43d1-9688-ff1bab454ba7 name=/runtime.v1.ImageService/ImageStatus Jan 05 08:07:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 08:07:36.444044938Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=4e9ec145-9d82-4350-ab57-a0812919bf23 name=/runtime.v1.ImageService/ImageStatus Jan 05 08:07:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 08:07:36.444318661Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=4e9ec145-9d82-4350-ab57-a0812919bf23 name=/runtime.v1.ImageService/ImageStatus Jan 05 08:12:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 08:12:36.446793290Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=7db5dfb1-7216-4eb6-8f05-0baad0b5c4e8 name=/runtime.v1.ImageService/ImageStatus Jan 05 08:12:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 08:12:36.447046584Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=7db5dfb1-7216-4eb6-8f05-0baad0b5c4e8 name=/runtime.v1.ImageService/ImageStatus Jan 05 08:17:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 08:17:36.449598882Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=0e5e12f6-33fa-4503-a47e-ce4528c875a2 name=/runtime.v1.ImageService/ImageStatus Jan 05 08:17:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 08:17:36.449850851Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=0e5e12f6-33fa-4503-a47e-ce4528c875a2 name=/runtime.v1.ImageService/ImageStatus Jan 05 08:22:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 08:22:36.452744603Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=534848d3-d841-4dc5-8a99-e8fe618598a1 name=/runtime.v1.ImageService/ImageStatus Jan 05 08:22:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 08:22:36.452989517Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=534848d3-d841-4dc5-8a99-e8fe618598a1 name=/runtime.v1.ImageService/ImageStatus Jan 05 08:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 08:27:36.455642338Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=b15f15af-f589-4416-8f1c-b667a179ab08 name=/runtime.v1.ImageService/ImageStatus Jan 05 08:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 08:27:36.455849548Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b15f15af-f589-4416-8f1c-b667a179ab08 name=/runtime.v1.ImageService/ImageStatus Jan 05 08:32:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 08:32:36.458225059Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=7aa0770d-259c-47d1-8920-6f4b09c4f1df name=/runtime.v1.ImageService/ImageStatus Jan 05 08:32:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 08:32:36.458473676Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=7aa0770d-259c-47d1-8920-6f4b09c4f1df name=/runtime.v1.ImageService/ImageStatus Jan 05 08:37:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 08:37:36.461235188Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=d2678351-f1a8-4bfe-bc74-9325fe3f3b17 name=/runtime.v1.ImageService/ImageStatus Jan 05 08:37:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 08:37:36.461489902Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=d2678351-f1a8-4bfe-bc74-9325fe3f3b17 name=/runtime.v1.ImageService/ImageStatus Jan 05 08:42:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 08:42:36.464756619Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=da1583e5-5dc4-44cd-bcd8-657bae57a2da name=/runtime.v1.ImageService/ImageStatus Jan 05 08:42:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 08:42:36.465015070Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=da1583e5-5dc4-44cd-bcd8-657bae57a2da name=/runtime.v1.ImageService/ImageStatus Jan 05 08:47:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 08:47:36.467578963Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=2f02bb68-9548-4f53-a965-2e8c2f7fa1fa name=/runtime.v1.ImageService/ImageStatus Jan 05 08:47:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 08:47:36.467797924Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=2f02bb68-9548-4f53-a965-2e8c2f7fa1fa name=/runtime.v1.ImageService/ImageStatus Jan 05 08:52:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 08:52:36.470563667Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=5dc707fe-0738-4eb3-9fd7-9f352eaa2eb2 name=/runtime.v1.ImageService/ImageStatus Jan 05 08:52:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 08:52:36.470816965Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=5dc707fe-0738-4eb3-9fd7-9f352eaa2eb2 name=/runtime.v1.ImageService/ImageStatus Jan 05 08:57:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 08:57:36.473946382Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=b6456df5-2d4c-4ce1-84d8-0f8ab5c50adf name=/runtime.v1.ImageService/ImageStatus Jan 05 08:57:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 08:57:36.474194924Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b6456df5-2d4c-4ce1-84d8-0f8ab5c50adf name=/runtime.v1.ImageService/ImageStatus Jan 05 09:02:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:02:36.476555685Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=f1528d35-14b4-4b63-9e1f-5dd327fdf5cd name=/runtime.v1.ImageService/ImageStatus Jan 05 09:02:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:02:36.476811043Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=f1528d35-14b4-4b63-9e1f-5dd327fdf5cd name=/runtime.v1.ImageService/ImageStatus Jan 05 09:07:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:07:36.480099378Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=eac4d827-73bb-4fe6-a1b9-ec0a05742532 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:07:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:07:36.480368317Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=eac4d827-73bb-4fe6-a1b9-ec0a05742532 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:11:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:27.633688781Z" level=warning msg="Found defunct process with PID 70768 (iptables)" Jan 05 09:11:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:29.757775 1425 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-kube-storage-version-migrator/migrator-6cd87bdff8-p9dkf] Jan 05 09:11:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:29.757829 1425 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:11:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:29.909081 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjtff\" (UniqueName: \"kubernetes.io/projected/26f1eb05-de00-4fc2-864c-6b97e7f82e96-kube-api-access-hjtff\") pod \"migrator-6cd87bdff8-p9dkf\" (UID: \"26f1eb05-de00-4fc2-864c-6b97e7f82e96\") " pod="openshift-kube-storage-version-migrator/migrator-6cd87bdff8-p9dkf" Jan 05 09:11:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:30.010225 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-hjtff\" (UniqueName: \"kubernetes.io/projected/26f1eb05-de00-4fc2-864c-6b97e7f82e96-kube-api-access-hjtff\") pod \"migrator-6cd87bdff8-p9dkf\" (UID: \"26f1eb05-de00-4fc2-864c-6b97e7f82e96\") " pod="openshift-kube-storage-version-migrator/migrator-6cd87bdff8-p9dkf" Jan 05 09:11:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:30.031072 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjtff\" (UniqueName: \"kubernetes.io/projected/26f1eb05-de00-4fc2-864c-6b97e7f82e96-kube-api-access-hjtff\") pod \"migrator-6cd87bdff8-p9dkf\" (UID: \"26f1eb05-de00-4fc2-864c-6b97e7f82e96\") " pod="openshift-kube-storage-version-migrator/migrator-6cd87bdff8-p9dkf" Jan 05 09:11:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:30.075636 1425 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-6cd87bdff8-p9dkf" Jan 05 09:11:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:30.076092572Z" level=info msg="Running pod sandbox: openshift-kube-storage-version-migrator/migrator-6cd87bdff8-p9dkf/POD" id=ceffdbc2-1586-493b-ae7c-781da17d123e name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:11:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:30.076154642Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:11:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:30.093974550Z" level=info msg="Got pod network &{Name:migrator-6cd87bdff8-p9dkf Namespace:openshift-kube-storage-version-migrator ID:148c6bc4da65fd9b07c5fca9d24c7096ee2b924f8d407958ac234e9758e340f4 UID:26f1eb05-de00-4fc2-864c-6b97e7f82e96 NetNS:/var/run/netns/7b49749d-6d7c-4bf9-bed4-8bb736f61214 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:11:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:30.094009945Z" level=info msg="Adding pod openshift-kube-storage-version-migrator_migrator-6cd87bdff8-p9dkf to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:11:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:32.162489 1425 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-kube-storage-version-migrator/migrator-6cd87bdff8-p9dkf] Jan 05 09:11:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: 2023-01-05T09:11:32Z [verbose] Add: openshift-kube-storage-version-migrator:migrator-6cd87bdff8-p9dkf:26f1eb05-de00-4fc2-864c-6b97e7f82e96:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/7b49749d-6d7c-4bf9-bed4-8bb736f61214"}],"ips":[{"version":"4","interface":0,"address":"10.129.2.6/23","gateway":"10.129.2.1"}],"dns":{}} Jan 05 09:11:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: I0105 09:11:32.138575 70899 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-kube-storage-version-migrator", Name:"migrator-6cd87bdff8-p9dkf", UID:"26f1eb05-de00-4fc2-864c-6b97e7f82e96", APIVersion:"v1", ResourceVersion:"73338", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.129.2.6/23] from openshift-sdn Jan 05 09:11:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:32.164733947Z" level=info msg="Got pod network &{Name:migrator-6cd87bdff8-p9dkf Namespace:openshift-kube-storage-version-migrator ID:148c6bc4da65fd9b07c5fca9d24c7096ee2b924f8d407958ac234e9758e340f4 UID:26f1eb05-de00-4fc2-864c-6b97e7f82e96 NetNS:/var/run/netns/7b49749d-6d7c-4bf9-bed4-8bb736f61214 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:11:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:32.164869507Z" level=info msg="Checking pod openshift-kube-storage-version-migrator_migrator-6cd87bdff8-p9dkf for CNI network multus-cni-network (type=multus)" Jan 05 09:11:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:11:32.166860 1425 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod26f1eb05_de00_4fc2_864c_6b97e7f82e96.slice/crio-148c6bc4da65fd9b07c5fca9d24c7096ee2b924f8d407958ac234e9758e340f4.scope WatchSource:0}: Error finding container 148c6bc4da65fd9b07c5fca9d24c7096ee2b924f8d407958ac234e9758e340f4: Status 404 returned error can't find the container with id 148c6bc4da65fd9b07c5fca9d24c7096ee2b924f8d407958ac234e9758e340f4 Jan 05 09:11:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:32.175235381Z" level=info msg="Ran pod sandbox 148c6bc4da65fd9b07c5fca9d24c7096ee2b924f8d407958ac234e9758e340f4 with infra container: openshift-kube-storage-version-migrator/migrator-6cd87bdff8-p9dkf/POD" id=ceffdbc2-1586-493b-ae7c-781da17d123e name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:11:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:32.176631145Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d0336c488437103651d61c3d74b49c13559c7cbc866106918883771b271d0ae9" id=6b788010-386c-49eb-ab33-edaaf32edd55 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:11:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:32.176819769Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d0336c488437103651d61c3d74b49c13559c7cbc866106918883771b271d0ae9 not found" id=6b788010-386c-49eb-ab33-edaaf32edd55 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:11:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:32.177133 1425 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 05 09:11:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:32.177522847Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d0336c488437103651d61c3d74b49c13559c7cbc866106918883771b271d0ae9" id=bd4fc555-7cce-4b5c-b572-dd5a7e32f4f3 name=/runtime.v1.ImageService/PullImage Jan 05 09:11:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:32.178438040Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d0336c488437103651d61c3d74b49c13559c7cbc866106918883771b271d0ae9\"" Jan 05 09:11:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:32.315539628Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d0336c488437103651d61c3d74b49c13559c7cbc866106918883771b271d0ae9\"" Jan 05 09:11:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:32.597353 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-6cd87bdff8-p9dkf" event=&{ID:26f1eb05-de00-4fc2-864c-6b97e7f82e96 Type:ContainerStarted Data:148c6bc4da65fd9b07c5fca9d24c7096ee2b924f8d407958ac234e9758e340f4} Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.150865 1425 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-image-registry/image-registry-64655f6865-dx2q6] Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.150931 1425 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.262972 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/4685a26b-f07f-4b10-8547-dea3d6802723-image-registry-private-configuration\") pod \"image-registry-64655f6865-dx2q6\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " pod="openshift-image-registry/image-registry-64655f6865-dx2q6" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.263056 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4685a26b-f07f-4b10-8547-dea3d6802723-ca-trust-extracted\") pod \"image-registry-64655f6865-dx2q6\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " pod="openshift-image-registry/image-registry-64655f6865-dx2q6" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.263097 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-db7jj\" (UniqueName: \"kubernetes.io/projected/4685a26b-f07f-4b10-8547-dea3d6802723-kube-api-access-db7jj\") pod \"image-registry-64655f6865-dx2q6\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " pod="openshift-image-registry/image-registry-64655f6865-dx2q6" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.263165 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4685a26b-f07f-4b10-8547-dea3d6802723-registry-tls\") pod \"image-registry-64655f6865-dx2q6\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " pod="openshift-image-registry/image-registry-64655f6865-dx2q6" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.263229 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4685a26b-f07f-4b10-8547-dea3d6802723-installation-pull-secrets\") pod \"image-registry-64655f6865-dx2q6\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " pod="openshift-image-registry/image-registry-64655f6865-dx2q6" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.263259 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4685a26b-f07f-4b10-8547-dea3d6802723-bound-sa-token\") pod \"image-registry-64655f6865-dx2q6\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " pod="openshift-image-registry/image-registry-64655f6865-dx2q6" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.263298 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4685a26b-f07f-4b10-8547-dea3d6802723-registry-certificates\") pod \"image-registry-64655f6865-dx2q6\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " pod="openshift-image-registry/image-registry-64655f6865-dx2q6" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.263349 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4685a26b-f07f-4b10-8547-dea3d6802723-trusted-ca\") pod \"image-registry-64655f6865-dx2q6\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " pod="openshift-image-registry/image-registry-64655f6865-dx2q6" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.364861 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/4685a26b-f07f-4b10-8547-dea3d6802723-image-registry-private-configuration\") pod \"image-registry-64655f6865-dx2q6\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " pod="openshift-image-registry/image-registry-64655f6865-dx2q6" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.364922 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4685a26b-f07f-4b10-8547-dea3d6802723-ca-trust-extracted\") pod \"image-registry-64655f6865-dx2q6\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " pod="openshift-image-registry/image-registry-64655f6865-dx2q6" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.364958 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-db7jj\" (UniqueName: \"kubernetes.io/projected/4685a26b-f07f-4b10-8547-dea3d6802723-kube-api-access-db7jj\") pod \"image-registry-64655f6865-dx2q6\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " pod="openshift-image-registry/image-registry-64655f6865-dx2q6" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.364997 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4685a26b-f07f-4b10-8547-dea3d6802723-registry-tls\") pod \"image-registry-64655f6865-dx2q6\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " pod="openshift-image-registry/image-registry-64655f6865-dx2q6" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.365061 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4685a26b-f07f-4b10-8547-dea3d6802723-installation-pull-secrets\") pod \"image-registry-64655f6865-dx2q6\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " pod="openshift-image-registry/image-registry-64655f6865-dx2q6" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.365094 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4685a26b-f07f-4b10-8547-dea3d6802723-bound-sa-token\") pod \"image-registry-64655f6865-dx2q6\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " pod="openshift-image-registry/image-registry-64655f6865-dx2q6" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.365133 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4685a26b-f07f-4b10-8547-dea3d6802723-registry-certificates\") pod \"image-registry-64655f6865-dx2q6\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " pod="openshift-image-registry/image-registry-64655f6865-dx2q6" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.365169 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4685a26b-f07f-4b10-8547-dea3d6802723-trusted-ca\") pod \"image-registry-64655f6865-dx2q6\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " pod="openshift-image-registry/image-registry-64655f6865-dx2q6" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.366253 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4685a26b-f07f-4b10-8547-dea3d6802723-trusted-ca\") pod \"image-registry-64655f6865-dx2q6\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " pod="openshift-image-registry/image-registry-64655f6865-dx2q6" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.368801 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4685a26b-f07f-4b10-8547-dea3d6802723-ca-trust-extracted\") pod \"image-registry-64655f6865-dx2q6\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " pod="openshift-image-registry/image-registry-64655f6865-dx2q6" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.370133 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4685a26b-f07f-4b10-8547-dea3d6802723-registry-certificates\") pod \"image-registry-64655f6865-dx2q6\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " pod="openshift-image-registry/image-registry-64655f6865-dx2q6" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.374387 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/4685a26b-f07f-4b10-8547-dea3d6802723-image-registry-private-configuration\") pod \"image-registry-64655f6865-dx2q6\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " pod="openshift-image-registry/image-registry-64655f6865-dx2q6" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.378998 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4685a26b-f07f-4b10-8547-dea3d6802723-registry-tls\") pod \"image-registry-64655f6865-dx2q6\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " pod="openshift-image-registry/image-registry-64655f6865-dx2q6" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.381173 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4685a26b-f07f-4b10-8547-dea3d6802723-installation-pull-secrets\") pod \"image-registry-64655f6865-dx2q6\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " pod="openshift-image-registry/image-registry-64655f6865-dx2q6" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.383712 1425 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-ingress/router-default-5cf6bbdff9-fkmcn] Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.383751 1425 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.409795 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-db7jj\" (UniqueName: \"kubernetes.io/projected/4685a26b-f07f-4b10-8547-dea3d6802723-kube-api-access-db7jj\") pod \"image-registry-64655f6865-dx2q6\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " pod="openshift-image-registry/image-registry-64655f6865-dx2q6" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.421482 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4685a26b-f07f-4b10-8547-dea3d6802723-bound-sa-token\") pod \"image-registry-64655f6865-dx2q6\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " pod="openshift-image-registry/image-registry-64655f6865-dx2q6" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.465701 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/66d22b70-08bb-4199-b258-ac4590e9b487-stats-auth\") pod \"router-default-5cf6bbdff9-fkmcn\" (UID: \"66d22b70-08bb-4199-b258-ac4590e9b487\") " pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.465770 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/66d22b70-08bb-4199-b258-ac4590e9b487-default-certificate\") pod \"router-default-5cf6bbdff9-fkmcn\" (UID: \"66d22b70-08bb-4199-b258-ac4590e9b487\") " pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.465805 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/66d22b70-08bb-4199-b258-ac4590e9b487-metrics-certs\") pod \"router-default-5cf6bbdff9-fkmcn\" (UID: \"66d22b70-08bb-4199-b258-ac4590e9b487\") " pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.465848 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66d22b70-08bb-4199-b258-ac4590e9b487-service-ca-bundle\") pod \"router-default-5cf6bbdff9-fkmcn\" (UID: \"66d22b70-08bb-4199-b258-ac4590e9b487\") " pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.465884 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzl7h\" (UniqueName: \"kubernetes.io/projected/66d22b70-08bb-4199-b258-ac4590e9b487-kube-api-access-vzl7h\") pod \"router-default-5cf6bbdff9-fkmcn\" (UID: \"66d22b70-08bb-4199-b258-ac4590e9b487\") " pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.474864 1425 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-64655f6865-dx2q6" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:34.475288400Z" level=info msg="Running pod sandbox: openshift-image-registry/image-registry-64655f6865-dx2q6/POD" id=873e35ff-52f0-439a-88b8-1554881989fb name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:34.475353128Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:34.523495775Z" level=info msg="Got pod network &{Name:image-registry-64655f6865-dx2q6 Namespace:openshift-image-registry ID:f855fd38c50954c79bc8f0f3e034b1d0f7293735e9dbc194b9e271cfb674c01c UID:4685a26b-f07f-4b10-8547-dea3d6802723 NetNS:/var/run/netns/c8c45494-0105-46cf-aadf-e33e0bf7b218 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:34.523539072Z" level=info msg="Adding pod openshift-image-registry_image-registry-64655f6865-dx2q6 to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.567590 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/66d22b70-08bb-4199-b258-ac4590e9b487-metrics-certs\") pod \"router-default-5cf6bbdff9-fkmcn\" (UID: \"66d22b70-08bb-4199-b258-ac4590e9b487\") " pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.567658 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66d22b70-08bb-4199-b258-ac4590e9b487-service-ca-bundle\") pod \"router-default-5cf6bbdff9-fkmcn\" (UID: \"66d22b70-08bb-4199-b258-ac4590e9b487\") " pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.567700 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-vzl7h\" (UniqueName: \"kubernetes.io/projected/66d22b70-08bb-4199-b258-ac4590e9b487-kube-api-access-vzl7h\") pod \"router-default-5cf6bbdff9-fkmcn\" (UID: \"66d22b70-08bb-4199-b258-ac4590e9b487\") " pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.567743 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/66d22b70-08bb-4199-b258-ac4590e9b487-stats-auth\") pod \"router-default-5cf6bbdff9-fkmcn\" (UID: \"66d22b70-08bb-4199-b258-ac4590e9b487\") " pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.567792 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/66d22b70-08bb-4199-b258-ac4590e9b487-default-certificate\") pod \"router-default-5cf6bbdff9-fkmcn\" (UID: \"66d22b70-08bb-4199-b258-ac4590e9b487\") " pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.569006 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66d22b70-08bb-4199-b258-ac4590e9b487-service-ca-bundle\") pod \"router-default-5cf6bbdff9-fkmcn\" (UID: \"66d22b70-08bb-4199-b258-ac4590e9b487\") " pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.575039 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/66d22b70-08bb-4199-b258-ac4590e9b487-default-certificate\") pod \"router-default-5cf6bbdff9-fkmcn\" (UID: \"66d22b70-08bb-4199-b258-ac4590e9b487\") " pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.576898 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/66d22b70-08bb-4199-b258-ac4590e9b487-metrics-certs\") pod \"router-default-5cf6bbdff9-fkmcn\" (UID: \"66d22b70-08bb-4199-b258-ac4590e9b487\") " pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.578505 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/66d22b70-08bb-4199-b258-ac4590e9b487-stats-auth\") pod \"router-default-5cf6bbdff9-fkmcn\" (UID: \"66d22b70-08bb-4199-b258-ac4590e9b487\") " pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.608197 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzl7h\" (UniqueName: \"kubernetes.io/projected/66d22b70-08bb-4199-b258-ac4590e9b487-kube-api-access-vzl7h\") pod \"router-default-5cf6bbdff9-fkmcn\" (UID: \"66d22b70-08bb-4199-b258-ac4590e9b487\") " pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.756812 1425 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:34.758216796Z" level=info msg="Running pod sandbox: openshift-ingress/router-default-5cf6bbdff9-fkmcn/POD" id=67cfb119-2f44-4b86-b506-44807c09ac3e name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:34.758271204Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.831357 1425 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-monitoring/prometheus-adapter-5894cc86c7-47422] Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.831424 1425 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.978338 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a82fd83b-5344-4e7a-91dd-5894f3846e87-serving-certs-ca-bundle\") pod \"prometheus-adapter-5894cc86c7-47422\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.978424 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a82fd83b-5344-4e7a-91dd-5894f3846e87-config\") pod \"prometheus-adapter-5894cc86c7-47422\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.978464 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-adapter-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/a82fd83b-5344-4e7a-91dd-5894f3846e87-prometheus-adapter-audit-profiles\") pod \"prometheus-adapter-5894cc86c7-47422\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.978504 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a82fd83b-5344-4e7a-91dd-5894f3846e87-tmpfs\") pod \"prometheus-adapter-5894cc86c7-47422\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.978551 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls\" (UniqueName: \"kubernetes.io/secret/a82fd83b-5344-4e7a-91dd-5894f3846e87-tls\") pod \"prometheus-adapter-5894cc86c7-47422\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.978600 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjb9d\" (UniqueName: \"kubernetes.io/projected/a82fd83b-5344-4e7a-91dd-5894f3846e87-kube-api-access-kjb9d\") pod \"prometheus-adapter-5894cc86c7-47422\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.978638 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-adapter-prometheus-config\" (UniqueName: \"kubernetes.io/configmap/a82fd83b-5344-4e7a-91dd-5894f3846e87-prometheus-adapter-prometheus-config\") pod \"prometheus-adapter-5894cc86c7-47422\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" Jan 05 09:11:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:34.978670 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/a82fd83b-5344-4e7a-91dd-5894f3846e87-audit-log\") pod \"prometheus-adapter-5894cc86c7-47422\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.082836 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"prometheus-adapter-prometheus-config\" (UniqueName: \"kubernetes.io/configmap/a82fd83b-5344-4e7a-91dd-5894f3846e87-prometheus-adapter-prometheus-config\") pod \"prometheus-adapter-5894cc86c7-47422\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.082900 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/a82fd83b-5344-4e7a-91dd-5894f3846e87-audit-log\") pod \"prometheus-adapter-5894cc86c7-47422\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.082949 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a82fd83b-5344-4e7a-91dd-5894f3846e87-serving-certs-ca-bundle\") pod \"prometheus-adapter-5894cc86c7-47422\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.082996 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a82fd83b-5344-4e7a-91dd-5894f3846e87-config\") pod \"prometheus-adapter-5894cc86c7-47422\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.083045 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"prometheus-adapter-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/a82fd83b-5344-4e7a-91dd-5894f3846e87-prometheus-adapter-audit-profiles\") pod \"prometheus-adapter-5894cc86c7-47422\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.083082 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a82fd83b-5344-4e7a-91dd-5894f3846e87-tmpfs\") pod \"prometheus-adapter-5894cc86c7-47422\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.083127 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tls\" (UniqueName: \"kubernetes.io/secret/a82fd83b-5344-4e7a-91dd-5894f3846e87-tls\") pod \"prometheus-adapter-5894cc86c7-47422\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.083165 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-kjb9d\" (UniqueName: \"kubernetes.io/projected/a82fd83b-5344-4e7a-91dd-5894f3846e87-kube-api-access-kjb9d\") pod \"prometheus-adapter-5894cc86c7-47422\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.084268 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"prometheus-adapter-prometheus-config\" (UniqueName: \"kubernetes.io/configmap/a82fd83b-5344-4e7a-91dd-5894f3846e87-prometheus-adapter-prometheus-config\") pod \"prometheus-adapter-5894cc86c7-47422\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.084645 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/a82fd83b-5344-4e7a-91dd-5894f3846e87-audit-log\") pod \"prometheus-adapter-5894cc86c7-47422\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.085188 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a82fd83b-5344-4e7a-91dd-5894f3846e87-serving-certs-ca-bundle\") pod \"prometheus-adapter-5894cc86c7-47422\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.085753 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a82fd83b-5344-4e7a-91dd-5894f3846e87-config\") pod \"prometheus-adapter-5894cc86c7-47422\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.086660 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"prometheus-adapter-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/a82fd83b-5344-4e7a-91dd-5894f3846e87-prometheus-adapter-audit-profiles\") pod \"prometheus-adapter-5894cc86c7-47422\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.086897 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a82fd83b-5344-4e7a-91dd-5894f3846e87-tmpfs\") pod \"prometheus-adapter-5894cc86c7-47422\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.089958 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tls\" (UniqueName: \"kubernetes.io/secret/a82fd83b-5344-4e7a-91dd-5894f3846e87-tls\") pod \"prometheus-adapter-5894cc86c7-47422\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.103188 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjb9d\" (UniqueName: \"kubernetes.io/projected/a82fd83b-5344-4e7a-91dd-5894f3846e87-kube-api-access-kjb9d\") pod \"prometheus-adapter-5894cc86c7-47422\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.143891 1425 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-54rjq] Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.143935 1425 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.167461 1425 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:35.175019980Z" level=info msg="Running pod sandbox: openshift-monitoring/prometheus-adapter-5894cc86c7-47422/POD" id=bfeb82c7-4bbe-4d4b-9b2f-8dba1b634a11 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:35.175082217Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.287088 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/f8e5be3f-9ab7-484c-b61b-6231a8774e24-tls-certificates\") pod \"prometheus-operator-admission-webhook-598468bb49-54rjq\" (UID: \"f8e5be3f-9ab7-484c-b61b-6231a8774e24\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-54rjq" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.393558 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/f8e5be3f-9ab7-484c-b61b-6231a8774e24-tls-certificates\") pod \"prometheus-operator-admission-webhook-598468bb49-54rjq\" (UID: \"f8e5be3f-9ab7-484c-b61b-6231a8774e24\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-54rjq" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.401695 1425 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx] Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.401749 1425 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.411010 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/f8e5be3f-9ab7-484c-b61b-6231a8774e24-tls-certificates\") pod \"prometheus-operator-admission-webhook-598468bb49-54rjq\" (UID: \"f8e5be3f-9ab7-484c-b61b-6231a8774e24\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-54rjq" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:35.412433325Z" level=info msg="Got pod network &{Name:prometheus-adapter-5894cc86c7-47422 Namespace:openshift-monitoring ID:ef7044b787198dfb2aab7533e9c72c07b8c2ef842ecb652a788dc647c485157c UID:a82fd83b-5344-4e7a-91dd-5894f3846e87 NetNS:/var/run/netns/5dddfea0-872a-4241-a44b-44a8fc6dedf6 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:35.412471895Z" level=info msg="Adding pod openshift-monitoring_prometheus-adapter-5894cc86c7-47422 to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:35.414133664Z" level=info msg="Got pod network &{Name:router-default-5cf6bbdff9-fkmcn Namespace:openshift-ingress ID:0412a26df3b1c530016b6935b36a365c8ab277d3b739d57e9cec85e5975d314c UID:66d22b70-08bb-4199-b258-ac4590e9b487 NetNS:/var/run/netns/c442ffc5-f1f5-46e0-870e-1ec985722737 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:35.414166279Z" level=info msg="Adding pod openshift-ingress_router-default-5cf6bbdff9-fkmcn to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.494942 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-querier-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-thanos-querier-trusted-ca-bundle\") pod \"thanos-querier-6f4f5fb7cf-wtcwx\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.495006 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-thanos-querier-tls\") pod \"thanos-querier-6f4f5fb7cf-wtcwx\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.495042 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-oauth-cookie\" (UniqueName: \"kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-thanos-querier-oauth-cookie\") pod \"thanos-querier-6f4f5fb7cf-wtcwx\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.495101 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-6f4f5fb7cf-wtcwx\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.495140 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brfpd\" (UniqueName: \"kubernetes.io/projected/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-kube-api-access-brfpd\") pod \"thanos-querier-6f4f5fb7cf-wtcwx\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.495177 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-grpc-tls\") pod \"thanos-querier-6f4f5fb7cf-wtcwx\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.495211 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-6f4f5fb7cf-wtcwx\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.495241 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-metrics-client-ca\") pod \"thanos-querier-6f4f5fb7cf-wtcwx\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.495279 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-6f4f5fb7cf-wtcwx\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.505743 1425 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-54rjq" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:35.506632766Z" level=info msg="Running pod sandbox: openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-54rjq/POD" id=32d03b52-25d8-4e4d-9448-4f5b6388fa5a name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:35.506703614Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.561819 1425 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-network-diagnostics/network-check-source-6775957967-zsr9c] Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.561867 1425 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:35.582082126Z" level=info msg="Got pod network &{Name:prometheus-operator-admission-webhook-598468bb49-54rjq Namespace:openshift-monitoring ID:8e21977f4f2aa77ecd7154452d4a15adc91d3b323c169dda297ad9c1f8901298 UID:f8e5be3f-9ab7-484c-b61b-6231a8774e24 NetNS:/var/run/netns/7ace1278-51dd-403a-b0d3-915a3ab684a5 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:35.582123746Z" level=info msg="Adding pod openshift-monitoring_prometheus-operator-admission-webhook-598468bb49-54rjq to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.596187 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-6f4f5fb7cf-wtcwx\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.596248 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-brfpd\" (UniqueName: \"kubernetes.io/projected/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-kube-api-access-brfpd\") pod \"thanos-querier-6f4f5fb7cf-wtcwx\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.596287 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-grpc-tls\") pod \"thanos-querier-6f4f5fb7cf-wtcwx\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.596325 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-6f4f5fb7cf-wtcwx\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.596359 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-metrics-client-ca\") pod \"thanos-querier-6f4f5fb7cf-wtcwx\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.596423 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-6f4f5fb7cf-wtcwx\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.596459 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"thanos-querier-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-thanos-querier-trusted-ca-bundle\") pod \"thanos-querier-6f4f5fb7cf-wtcwx\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.596504 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-thanos-querier-tls\") pod \"thanos-querier-6f4f5fb7cf-wtcwx\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.596554 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-oauth-cookie\" (UniqueName: \"kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-thanos-querier-oauth-cookie\") pod \"thanos-querier-6f4f5fb7cf-wtcwx\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.599047 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-metrics-client-ca\") pod \"thanos-querier-6f4f5fb7cf-wtcwx\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.599862 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"thanos-querier-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-thanos-querier-trusted-ca-bundle\") pod \"thanos-querier-6f4f5fb7cf-wtcwx\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.608163 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-oauth-cookie\" (UniqueName: \"kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-thanos-querier-oauth-cookie\") pod \"thanos-querier-6f4f5fb7cf-wtcwx\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.611891 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-grpc-tls\") pod \"thanos-querier-6f4f5fb7cf-wtcwx\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.614183 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-6f4f5fb7cf-wtcwx\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.617275 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-6f4f5fb7cf-wtcwx\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.641152 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-6f4f5fb7cf-wtcwx\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.643739 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-thanos-querier-tls\") pod \"thanos-querier-6f4f5fb7cf-wtcwx\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.673887 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-brfpd\" (UniqueName: \"kubernetes.io/projected/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-kube-api-access-brfpd\") pod \"thanos-querier-6f4f5fb7cf-wtcwx\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.698915 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4df76\" (UniqueName: \"kubernetes.io/projected/89a48689-810c-464e-9ac6-d78977e60793-kube-api-access-4df76\") pod \"network-check-source-6775957967-zsr9c\" (UID: \"89a48689-810c-464e-9ac6-d78977e60793\") " pod="openshift-network-diagnostics/network-check-source-6775957967-zsr9c" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.800188 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-4df76\" (UniqueName: \"kubernetes.io/projected/89a48689-810c-464e-9ac6-d78977e60793-kube-api-access-4df76\") pod \"network-check-source-6775957967-zsr9c\" (UID: \"89a48689-810c-464e-9ac6-d78977e60793\") " pod="openshift-network-diagnostics/network-check-source-6775957967-zsr9c" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.833401 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-4df76\" (UniqueName: \"kubernetes.io/projected/89a48689-810c-464e-9ac6-d78977e60793-kube-api-access-4df76\") pod \"network-check-source-6775957967-zsr9c\" (UID: \"89a48689-810c-464e-9ac6-d78977e60793\") " pod="openshift-network-diagnostics/network-check-source-6775957967-zsr9c" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:35.841879 1425 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:35.846745969Z" level=info msg="Running pod sandbox: openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/POD" id=62fcd0f6-9d9d-471f-a058-c23e9c2d03fa name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:35.846806708Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:35.877425163Z" level=info msg="Got pod network &{Name:thanos-querier-6f4f5fb7cf-wtcwx Namespace:openshift-monitoring ID:6c2cdef5be97661e24b16499c5f84baa5f0fdffbceb45f19b515c50e93bb6dd5 UID:72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c NetNS:/var/run/netns/0dfe86e0-4e47-48fc-98f3-38c9aac3ee85 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:11:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:35.877463798Z" level=info msg="Adding pod openshift-monitoring_thanos-querier-6f4f5fb7cf-wtcwx to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:11:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:36.037470 1425 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-6775957967-zsr9c" Jan 05 09:11:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:36.039537099Z" level=info msg="Running pod sandbox: openshift-network-diagnostics/network-check-source-6775957967-zsr9c/POD" id=bfb7fe3a-8128-4d0a-8f82-88e7ec42edcd name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:11:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:36.039624299Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:11:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:36.063100580Z" level=info msg="Got pod network &{Name:network-check-source-6775957967-zsr9c Namespace:openshift-network-diagnostics ID:e7532666db14128bc848af4cbb52adc603db1ec585b807063b42799df403a10b UID:89a48689-810c-464e-9ac6-d78977e60793 NetNS:/var/run/netns/61898809-3b8f-4a3e-bbd7-8989e7ede22e Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:11:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:36.063135944Z" level=info msg="Adding pod openshift-network-diagnostics_network-check-source-6775957967-zsr9c to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:11:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: 2023-01-05T09:11:36Z [verbose] Add: openshift-image-registry:image-registry-64655f6865-dx2q6:4685a26b-f07f-4b10-8547-dea3d6802723:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/c8c45494-0105-46cf-aadf-e33e0bf7b218"}],"ips":[{"version":"4","interface":0,"address":"10.129.2.7/23","gateway":"10.129.2.1"}],"dns":{}} Jan 05 09:11:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: I0105 09:11:36.157268 71130 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-image-registry", Name:"image-registry-64655f6865-dx2q6", UID:"4685a26b-f07f-4b10-8547-dea3d6802723", APIVersion:"v1", ResourceVersion:"73597", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.129.2.7/23] from openshift-sdn Jan 05 09:11:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:36.178442565Z" level=info msg="Got pod network &{Name:image-registry-64655f6865-dx2q6 Namespace:openshift-image-registry ID:f855fd38c50954c79bc8f0f3e034b1d0f7293735e9dbc194b9e271cfb674c01c UID:4685a26b-f07f-4b10-8547-dea3d6802723 NetNS:/var/run/netns/c8c45494-0105-46cf-aadf-e33e0bf7b218 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:11:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:36.178583692Z" level=info msg="Checking pod openshift-image-registry_image-registry-64655f6865-dx2q6 for CNI network multus-cni-network (type=multus)" Jan 05 09:11:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:36.178091 1425 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-image-registry/image-registry-64655f6865-dx2q6] Jan 05 09:11:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:11:36.180595 1425 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4685a26b_f07f_4b10_8547_dea3d6802723.slice/crio-f855fd38c50954c79bc8f0f3e034b1d0f7293735e9dbc194b9e271cfb674c01c.scope WatchSource:0}: Error finding container f855fd38c50954c79bc8f0f3e034b1d0f7293735e9dbc194b9e271cfb674c01c: Status 404 returned error can't find the container with id f855fd38c50954c79bc8f0f3e034b1d0f7293735e9dbc194b9e271cfb674c01c Jan 05 09:11:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:36.187918830Z" level=info msg="Ran pod sandbox f855fd38c50954c79bc8f0f3e034b1d0f7293735e9dbc194b9e271cfb674c01c with infra container: openshift-image-registry/image-registry-64655f6865-dx2q6/POD" id=873e35ff-52f0-439a-88b8-1554881989fb name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:11:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:36.190749374Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8" id=6f336219-da87-435a-becb-b3c28e0fb15b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:11:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:36.190958187Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:c9e2954d8b9d9e1594d0a16b90086a9297ffd65773a6aa4a64f325e871ce83d1,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8],Size_:363009057,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=6f336219-da87-435a-becb-b3c28e0fb15b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:11:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:36.192097242Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8" id=785925ac-dffa-40f4-a3f0-405a0997ac4e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:11:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:36.192280177Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:c9e2954d8b9d9e1594d0a16b90086a9297ffd65773a6aa4a64f325e871ce83d1,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8],Size_:363009057,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=785925ac-dffa-40f4-a3f0-405a0997ac4e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:11:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:36.193883083Z" level=info msg="Creating container: openshift-image-registry/image-registry-64655f6865-dx2q6/registry" id=c272b2b5-e01f-4cbf-b83f-bc58b7373440 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:11:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:36.193997462Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:11:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:36.390122108Z" level=info msg="Created container 784e9766ec3aebd02560a716704c4d48760309927d4008e2c136d460249ddb98: openshift-image-registry/image-registry-64655f6865-dx2q6/registry" id=c272b2b5-e01f-4cbf-b83f-bc58b7373440 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:11:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:36.390770671Z" level=info msg="Starting container: 784e9766ec3aebd02560a716704c4d48760309927d4008e2c136d460249ddb98" id=9fe84c55-fc98-485a-b580-567a53fbc346 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:11:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:36.431077651Z" level=info msg="Started container" PID=71388 containerID=784e9766ec3aebd02560a716704c4d48760309927d4008e2c136d460249ddb98 description=openshift-image-registry/image-registry-64655f6865-dx2q6/registry id=9fe84c55-fc98-485a-b580-567a53fbc346 name=/runtime.v1.RuntimeService/StartContainer sandboxID=f855fd38c50954c79bc8f0f3e034b1d0f7293735e9dbc194b9e271cfb674c01c Jan 05 09:11:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:36.719020 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-64655f6865-dx2q6" event=&{ID:4685a26b-f07f-4b10-8547-dea3d6802723 Type:ContainerStarted Data:784e9766ec3aebd02560a716704c4d48760309927d4008e2c136d460249ddb98} Jan 05 09:11:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:36.719056 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-64655f6865-dx2q6" event=&{ID:4685a26b-f07f-4b10-8547-dea3d6802723 Type:ContainerStarted Data:f855fd38c50954c79bc8f0f3e034b1d0f7293735e9dbc194b9e271cfb674c01c} Jan 05 09:11:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:36.720431 1425 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-64655f6865-dx2q6" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.030697 1425 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-ingress/router-default-5cf6bbdff9-fkmcn] Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: 2023-01-05T09:11:37Z [verbose] Add: openshift-ingress:router-default-5cf6bbdff9-fkmcn:66d22b70-08bb-4199-b258-ac4590e9b487:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/c442ffc5-f1f5-46e0-870e-1ec985722737"}],"ips":[{"version":"4","interface":0,"address":"10.129.2.8/23","gateway":"10.129.2.1"}],"dns":{}} Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: I0105 09:11:37.015035 71195 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-ingress", Name:"router-default-5cf6bbdff9-fkmcn", UID:"66d22b70-08bb-4199-b258-ac4590e9b487", APIVersion:"v1", ResourceVersion:"73624", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.129.2.8/23] from openshift-sdn Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.031773823Z" level=info msg="Got pod network &{Name:router-default-5cf6bbdff9-fkmcn Namespace:openshift-ingress ID:0412a26df3b1c530016b6935b36a365c8ab277d3b739d57e9cec85e5975d314c UID:66d22b70-08bb-4199-b258-ac4590e9b487 NetNS:/var/run/netns/c442ffc5-f1f5-46e0-870e-1ec985722737 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.031918009Z" level=info msg="Checking pod openshift-ingress_router-default-5cf6bbdff9-fkmcn for CNI network multus-cni-network (type=multus)" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:11:37.033890 1425 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66d22b70_08bb_4199_b258_ac4590e9b487.slice/crio-0412a26df3b1c530016b6935b36a365c8ab277d3b739d57e9cec85e5975d314c.scope WatchSource:0}: Error finding container 0412a26df3b1c530016b6935b36a365c8ab277d3b739d57e9cec85e5975d314c: Status 404 returned error can't find the container with id 0412a26df3b1c530016b6935b36a365c8ab277d3b739d57e9cec85e5975d314c Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.040060519Z" level=info msg="Ran pod sandbox 0412a26df3b1c530016b6935b36a365c8ab277d3b739d57e9cec85e5975d314c with infra container: openshift-ingress/router-default-5cf6bbdff9-fkmcn/POD" id=67cfb119-2f44-4b86-b506-44807c09ac3e name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.042158943Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:bfd14188641a2d34e5b2b2ed3336b3b440155533f3fd7c6de776300dc3ff9e3b" id=52b6097a-313f-45c8-ad54-21e8c1fbc9d7 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.042345017Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:bfd14188641a2d34e5b2b2ed3336b3b440155533f3fd7c6de776300dc3ff9e3b not found" id=52b6097a-313f-45c8-ad54-21e8c1fbc9d7 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.043005829Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:bfd14188641a2d34e5b2b2ed3336b3b440155533f3fd7c6de776300dc3ff9e3b" id=477433b0-3246-4ca0-adf1-eee7b7dbd563 name=/runtime.v1.ImageService/PullImage Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.043874006Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:bfd14188641a2d34e5b2b2ed3336b3b440155533f3fd7c6de776300dc3ff9e3b\"" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.314381 1425 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/prometheus-adapter-5894cc86c7-47422] Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: 2023-01-05T09:11:37Z [verbose] Add: openshift-monitoring:prometheus-adapter-5894cc86c7-47422:a82fd83b-5344-4e7a-91dd-5894f3846e87:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/5dddfea0-872a-4241-a44b-44a8fc6dedf6"}],"ips":[{"version":"4","interface":0,"address":"10.129.2.9/23","gateway":"10.129.2.1"}],"dns":{}} Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: I0105 09:11:37.292114 71199 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"prometheus-adapter-5894cc86c7-47422", UID:"a82fd83b-5344-4e7a-91dd-5894f3846e87", APIVersion:"v1", ResourceVersion:"73671", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.129.2.9/23] from openshift-sdn Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.320890922Z" level=info msg="Got pod network &{Name:prometheus-adapter-5894cc86c7-47422 Namespace:openshift-monitoring ID:ef7044b787198dfb2aab7533e9c72c07b8c2ef842ecb652a788dc647c485157c UID:a82fd83b-5344-4e7a-91dd-5894f3846e87 NetNS:/var/run/netns/5dddfea0-872a-4241-a44b-44a8fc6dedf6 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.321100995Z" level=info msg="Checking pod openshift-monitoring_prometheus-adapter-5894cc86c7-47422 for CNI network multus-cni-network (type=multus)" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.336033842Z" level=info msg="Ran pod sandbox ef7044b787198dfb2aab7533e9c72c07b8c2ef842ecb652a788dc647c485157c with infra container: openshift-monitoring/prometheus-adapter-5894cc86c7-47422/POD" id=bfeb82c7-4bbe-4d4b-9b2f-8dba1b634a11 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:11:37.337018 1425 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda82fd83b_5344_4e7a_91dd_5894f3846e87.slice/crio-ef7044b787198dfb2aab7533e9c72c07b8c2ef842ecb652a788dc647c485157c.scope WatchSource:0}: Error finding container ef7044b787198dfb2aab7533e9c72c07b8c2ef842ecb652a788dc647c485157c: Status 404 returned error can't find the container with id ef7044b787198dfb2aab7533e9c72c07b8c2ef842ecb652a788dc647c485157c Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.338571600Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9696f8f231d49aadf63ab980ec5c17c4fd87011930daecd30cc9e4ee102f1aeb" id=3d896969-7571-4dec-8ce5-aad2bc7cd8dd name=/runtime.v1.ImageService/ImageStatus Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.338794587Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9696f8f231d49aadf63ab980ec5c17c4fd87011930daecd30cc9e4ee102f1aeb not found" id=3d896969-7571-4dec-8ce5-aad2bc7cd8dd name=/runtime.v1.ImageService/ImageStatus Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.339722905Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9696f8f231d49aadf63ab980ec5c17c4fd87011930daecd30cc9e4ee102f1aeb" id=9fb1448d-d561-4c4e-8bed-b69ea7bc0ee6 name=/runtime.v1.ImageService/PullImage Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.340807256Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9696f8f231d49aadf63ab980ec5c17c4fd87011930daecd30cc9e4ee102f1aeb\"" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.502610541Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9696f8f231d49aadf63ab980ec5c17c4fd87011930daecd30cc9e4ee102f1aeb\"" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.537113 1425 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx] Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: 2023-01-05T09:11:37Z [verbose] Add: openshift-monitoring:thanos-querier-6f4f5fb7cf-wtcwx:72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/0dfe86e0-4e47-48fc-98f3-38c9aac3ee85"}],"ips":[{"version":"4","interface":0,"address":"10.129.2.11/23","gateway":"10.129.2.1"}],"dns":{}} Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: I0105 09:11:37.496454 71298 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"thanos-querier-6f4f5fb7cf-wtcwx", UID:"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c", APIVersion:"v1", ResourceVersion:"73735", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.129.2.11/23] from openshift-sdn Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.541540696Z" level=info msg="Got pod network &{Name:thanos-querier-6f4f5fb7cf-wtcwx Namespace:openshift-monitoring ID:6c2cdef5be97661e24b16499c5f84baa5f0fdffbceb45f19b515c50e93bb6dd5 UID:72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c NetNS:/var/run/netns/0dfe86e0-4e47-48fc-98f3-38c9aac3ee85 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.541723630Z" level=info msg="Checking pod openshift-monitoring_thanos-querier-6f4f5fb7cf-wtcwx for CNI network multus-cni-network (type=multus)" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.544438 1425 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-monitoring/alertmanager-main-1] Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.544492 1425 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.551850373Z" level=info msg="Ran pod sandbox 6c2cdef5be97661e24b16499c5f84baa5f0fdffbceb45f19b515c50e93bb6dd5 with infra container: openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/POD" id=62fcd0f6-9d9d-471f-a058-c23e9c2d03fa name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.553916224Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc" id=2bfca66e-f739-47e0-aff0-ef80bcd71b36 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.554125626Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc not found" id=2bfca66e-f739-47e0-aff0-ef80bcd71b36 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.564456683Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc" id=c5819c43-72d5-457a-92f3-f99100961920 name=/runtime.v1.ImageService/PullImage Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.565494416Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc\"" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.592092 1425 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-54rjq] Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: 2023-01-05T09:11:37Z [verbose] Add: openshift-monitoring:prometheus-operator-admission-webhook-598468bb49-54rjq:f8e5be3f-9ab7-484c-b61b-6231a8774e24:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/7ace1278-51dd-403a-b0d3-915a3ab684a5"}],"ips":[{"version":"4","interface":0,"address":"10.129.2.10/23","gateway":"10.129.2.1"}],"dns":{}} Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: I0105 09:11:37.574499 71252 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"prometheus-operator-admission-webhook-598468bb49-54rjq", UID:"f8e5be3f-9ab7-484c-b61b-6231a8774e24", APIVersion:"v1", ResourceVersion:"73713", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.129.2.10/23] from openshift-sdn Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.595319440Z" level=info msg="Got pod network &{Name:prometheus-operator-admission-webhook-598468bb49-54rjq Namespace:openshift-monitoring ID:8e21977f4f2aa77ecd7154452d4a15adc91d3b323c169dda297ad9c1f8901298 UID:f8e5be3f-9ab7-484c-b61b-6231a8774e24 NetNS:/var/run/netns/7ace1278-51dd-403a-b0d3-915a3ab684a5 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.595499665Z" level=info msg="Checking pod openshift-monitoring_prometheus-operator-admission-webhook-598468bb49-54rjq for CNI network multus-cni-network (type=multus)" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.599223346Z" level=info msg="Ran pod sandbox 8e21977f4f2aa77ecd7154452d4a15adc91d3b323c169dda297ad9c1f8901298 with infra container: openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-54rjq/POD" id=32d03b52-25d8-4e4d-9448-4f5b6388fa5a name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.608586257Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:dbf46164dcc3b2b6ef5df5d043ddc4fd859fd2c73c4862808424b3df6a1e27e7" id=f8d79dce-0c0c-4ad4-9cae-4c1463853237 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.608815310Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:dbf46164dcc3b2b6ef5df5d043ddc4fd859fd2c73c4862808424b3df6a1e27e7 not found" id=f8d79dce-0c0c-4ad4-9cae-4c1463853237 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.609902784Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:dbf46164dcc3b2b6ef5df5d043ddc4fd859fd2c73c4862808424b3df6a1e27e7" id=67f26b37-1aa8-4f5e-83fc-9922fe8fb8b8 name=/runtime.v1.ImageService/PullImage Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.610774419Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:dbf46164dcc3b2b6ef5df5d043ddc4fd859fd2c73c4862808424b3df6a1e27e7\"" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.630862 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-proxy\" (UniqueName: \"kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-secret-alertmanager-main-proxy\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.630908 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b59f4808-02df-4b2e-9694-8cf72a1b5e75-tls-assets\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.630944 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.630979 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-config-volume\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.631014 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.631060 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-secret-alertmanager-main-tls\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.631149 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqgp9\" (UniqueName: \"kubernetes.io/projected/b59f4808-02df-4b2e-9694-8cf72a1b5e75-kube-api-access-xqgp9\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.631193 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/b59f4808-02df-4b2e-9694-8cf72a1b5e75-alertmanager-main-db\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.631249 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-web-config\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.631393 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/b59f4808-02df-4b2e-9694-8cf72a1b5e75-metrics-client-ca\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.631462 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b59f4808-02df-4b2e-9694-8cf72a1b5e75-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.631501 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b59f4808-02df-4b2e-9694-8cf72a1b5e75-config-out\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.678704 1425 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-network-diagnostics/network-check-source-6775957967-zsr9c] Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: 2023-01-05T09:11:37Z [verbose] Add: openshift-network-diagnostics:network-check-source-6775957967-zsr9c:89a48689-810c-464e-9ac6-d78977e60793:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/61898809-3b8f-4a3e-bbd7-8989e7ede22e"}],"ips":[{"version":"4","interface":0,"address":"10.129.2.12/23","gateway":"10.129.2.1"}],"dns":{}} Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: I0105 09:11:37.660647 71331 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-network-diagnostics", Name:"network-check-source-6775957967-zsr9c", UID:"89a48689-810c-464e-9ac6-d78977e60793", APIVersion:"v1", ResourceVersion:"73750", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.129.2.12/23] from openshift-sdn Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.681967849Z" level=info msg="Got pod network &{Name:network-check-source-6775957967-zsr9c Namespace:openshift-network-diagnostics ID:e7532666db14128bc848af4cbb52adc603db1ec585b807063b42799df403a10b UID:89a48689-810c-464e-9ac6-d78977e60793 NetNS:/var/run/netns/61898809-3b8f-4a3e-bbd7-8989e7ede22e Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.682102849Z" level=info msg="Checking pod openshift-network-diagnostics_network-check-source-6775957967-zsr9c for CNI network multus-cni-network (type=multus)" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:11:37.690234 1425 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89a48689_810c_464e_9ac6_d78977e60793.slice/crio-e7532666db14128bc848af4cbb52adc603db1ec585b807063b42799df403a10b.scope WatchSource:0}: Error finding container e7532666db14128bc848af4cbb52adc603db1ec585b807063b42799df403a10b: Status 404 returned error can't find the container with id e7532666db14128bc848af4cbb52adc603db1ec585b807063b42799df403a10b Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.694034558Z" level=info msg="Ran pod sandbox e7532666db14128bc848af4cbb52adc603db1ec585b807063b42799df403a10b with infra container: openshift-network-diagnostics/network-check-source-6775957967-zsr9c/POD" id=bfb7fe3a-8128-4d0a-8f82-88e7ec42edcd name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.694886228Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7" id=8f15b083-7894-4a68-8f32-7ec8668688b2 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.695091005Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:4cb41c73a232ec4401c5db32cdb523810dc12ac605cdc6a9ce388e90930eb6c4,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7],Size_:515955107,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=8f15b083-7894-4a68-8f32-7ec8668688b2 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.695746688Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7" id=dcd381fe-62cc-44ad-945a-6fc23bbf22b7 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.695921112Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:4cb41c73a232ec4401c5db32cdb523810dc12ac605cdc6a9ce388e90930eb6c4,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7],Size_:515955107,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=dcd381fe-62cc-44ad-945a-6fc23bbf22b7 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.696577777Z" level=info msg="Creating container: openshift-network-diagnostics/network-check-source-6775957967-zsr9c/check-endpoints" id=32932d1c-f8ea-44b6-b172-728c5380fee3 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.696685161Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.723781 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-54rjq" event=&{ID:f8e5be3f-9ab7-484c-b61b-6231a8774e24 Type:ContainerStarted Data:8e21977f4f2aa77ecd7154452d4a15adc91d3b323c169dda297ad9c1f8901298} Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.731454 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-6775957967-zsr9c" event=&{ID:89a48689-810c-464e-9ac6-d78977e60793 Type:ContainerStarted Data:e7532666db14128bc848af4cbb52adc603db1ec585b807063b42799df403a10b} Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.734841 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b59f4808-02df-4b2e-9694-8cf72a1b5e75-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.734891 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b59f4808-02df-4b2e-9694-8cf72a1b5e75-config-out\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.734934 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-proxy\" (UniqueName: \"kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-secret-alertmanager-main-proxy\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.734968 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b59f4808-02df-4b2e-9694-8cf72a1b5e75-tls-assets\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.735002 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.735035 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-config-volume\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.735067 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.735105 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-secret-alertmanager-main-tls\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.735136 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-xqgp9\" (UniqueName: \"kubernetes.io/projected/b59f4808-02df-4b2e-9694-8cf72a1b5e75-kube-api-access-xqgp9\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.735176 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/b59f4808-02df-4b2e-9694-8cf72a1b5e75-alertmanager-main-db\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.735207 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-web-config\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.735244 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/b59f4808-02df-4b2e-9694-8cf72a1b5e75-metrics-client-ca\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.735946 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/b59f4808-02df-4b2e-9694-8cf72a1b5e75-metrics-client-ca\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.736855 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b59f4808-02df-4b2e-9694-8cf72a1b5e75-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.737143 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b59f4808-02df-4b2e-9694-8cf72a1b5e75-config-out\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.739626 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/b59f4808-02df-4b2e-9694-8cf72a1b5e75-alertmanager-main-db\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.742661 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-proxy\" (UniqueName: \"kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-secret-alertmanager-main-proxy\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.744855 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.746925 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b59f4808-02df-4b2e-9694-8cf72a1b5e75-tls-assets\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.749045 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.751633126Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc\"" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.752116 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-config-volume\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.752534 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" event=&{ID:72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c Type:ContainerStarted Data:6c2cdef5be97661e24b16499c5f84baa5f0fdffbceb45f19b515c50e93bb6dd5} Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.755860 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-web-config\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.756608 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-secret-alertmanager-main-tls\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.761180 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" event=&{ID:66d22b70-08bb-4199-b258-ac4590e9b487 Type:ContainerStarted Data:0412a26df3b1c530016b6935b36a365c8ab277d3b739d57e9cec85e5975d314c} Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.763617 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqgp9\" (UniqueName: \"kubernetes.io/projected/b59f4808-02df-4b2e-9694-8cf72a1b5e75-kube-api-access-xqgp9\") pod \"alertmanager-main-1\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.778090 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" event=&{ID:a82fd83b-5344-4e7a-91dd-5894f3846e87 Type:ContainerStarted Data:ef7044b787198dfb2aab7533e9c72c07b8c2ef842ecb652a788dc647c485157c} Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.874155625Z" level=info msg="Created container 3b8c93b35afc863ee93d43a145f66340ca0ab0a562aca1259bff690246d320cf: openshift-network-diagnostics/network-check-source-6775957967-zsr9c/check-endpoints" id=32932d1c-f8ea-44b6-b172-728c5380fee3 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.874774394Z" level=info msg="Starting container: 3b8c93b35afc863ee93d43a145f66340ca0ab0a562aca1259bff690246d320cf" id=cb5cce0d-311c-430a-9f37-a303bd6698df name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.884037057Z" level=info msg="Started container" PID=71578 containerID=3b8c93b35afc863ee93d43a145f66340ca0ab0a562aca1259bff690246d320cf description=openshift-network-diagnostics/network-check-source-6775957967-zsr9c/check-endpoints id=cb5cce0d-311c-430a-9f37-a303bd6698df name=/runtime.v1.RuntimeService/StartContainer sandboxID=e7532666db14128bc848af4cbb52adc603db1ec585b807063b42799df403a10b Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:37.900947 1425 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.903300253Z" level=info msg="Running pod sandbox: openshift-monitoring/alertmanager-main-1/POD" id=32bd009c-ff02-4645-af29-df6fcc619622 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.903351855Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.938237270Z" level=info msg="Got pod network &{Name:alertmanager-main-1 Namespace:openshift-monitoring ID:4cc6b8c0009aed4f5500066da7f716df203dd1969281bb8dbce4d7052e55ec50 UID:b59f4808-02df-4b2e-9694-8cf72a1b5e75 NetNS:/var/run/netns/a22d0db9-a46e-42fb-931a-4c5793845fd1 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:11:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:37.938272664Z" level=info msg="Adding pod openshift-monitoring_alertmanager-main-1 to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:38.218967772Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:bfd14188641a2d34e5b2b2ed3336b3b440155533f3fd7c6de776300dc3ff9e3b\"" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.500538 1425 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-monitoring/prometheus-k8s-1] Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.500596 1425 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.665911 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-kube-etcd-client-certs\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-kube-etcd-client-certs\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.665964 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-metrics-client-ca\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.666005 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8ac551e8-a959-4212-86a1-2b3ac482586f-tls-assets\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.666040 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.666082 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99gqh\" (UniqueName: \"kubernetes.io/projected/8ac551e8-a959-4212-86a1-2b3ac482586f-kube-api-access-99gqh\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.666118 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.666149 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-web-config\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.666182 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.666217 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.666251 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-config\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.666286 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-configmap-metrics-client-ca\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.666317 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.666348 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/8ac551e8-a959-4212-86a1-2b3ac482586f-prometheus-k8s-db\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.666401 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-proxy\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-prometheus-k8s-proxy\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.666432 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8ac551e8-a959-4212-86a1-2b3ac482586f-config-out\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.666464 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-kube-rbac-proxy\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.666497 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.666527 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-grpc-tls\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.666570 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-metrics-client-certs\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:38.739649977Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:dbf46164dcc3b2b6ef5df5d043ddc4fd859fd2c73c4862808424b3df6a1e27e7\"" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.767069 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.767129 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-config\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.767185 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-configmap-metrics-client-ca\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.767235 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.767274 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/8ac551e8-a959-4212-86a1-2b3ac482586f-prometheus-k8s-db\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.767306 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-proxy\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-prometheus-k8s-proxy\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.767336 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8ac551e8-a959-4212-86a1-2b3ac482586f-config-out\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.767387 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-kube-rbac-proxy\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.767424 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.767456 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-grpc-tls\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.767518 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-metrics-client-certs\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.767572 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-kube-etcd-client-certs\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-kube-etcd-client-certs\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.767605 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-metrics-client-ca\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.767642 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8ac551e8-a959-4212-86a1-2b3ac482586f-tls-assets\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.767674 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.767734 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.767767 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-99gqh\" (UniqueName: \"kubernetes.io/projected/8ac551e8-a959-4212-86a1-2b3ac482586f-kube-api-access-99gqh\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.767797 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-web-config\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.767831 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.772903 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.774628 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-configmap-metrics-client-ca\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.775401 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.775701 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/8ac551e8-a959-4212-86a1-2b3ac482586f-prometheus-k8s-db\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.776167 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8ac551e8-a959-4212-86a1-2b3ac482586f-config-out\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.779773 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-metrics-client-ca\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.780384 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.781044 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.784093 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.786346 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-grpc-tls\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.788470 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-config\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.791397 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-proxy\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-prometheus-k8s-proxy\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.808433 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-kube-rbac-proxy\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.812832 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-6775957967-zsr9c" event=&{ID:89a48689-810c-464e-9ac6-d78977e60793 Type:ContainerStarted Data:3b8c93b35afc863ee93d43a145f66340ca0ab0a562aca1259bff690246d320cf} Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.816071 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.819582 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8ac551e8-a959-4212-86a1-2b3ac482586f-tls-assets\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.822556 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-metrics-client-certs\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.825076 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-kube-etcd-client-certs\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-kube-etcd-client-certs\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.827563 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-web-config\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.834834 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-99gqh\" (UniqueName: \"kubernetes.io/projected/8ac551e8-a959-4212-86a1-2b3ac482586f-kube-api-access-99gqh\") pod \"prometheus-k8s-1\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:38.918148 1425 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:38.919307749Z" level=info msg="Running pod sandbox: openshift-monitoring/prometheus-k8s-1/POD" id=326dcfe0-3314-4230-b163-08231b9e8df1 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:38.919386092Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:38.986945959Z" level=info msg="Got pod network &{Name:prometheus-k8s-1 Namespace:openshift-monitoring ID:bf998995ead8afb2642821b82636d7b867d8c0aa5ab164fe09e7396c2701f63b UID:8ac551e8-a959-4212-86a1-2b3ac482586f NetNS:/var/run/netns/cf32a4f0-a7ea-4d0a-ada0-ddff6598e5f3 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:11:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:38.986982738Z" level=info msg="Adding pod openshift-monitoring_prometheus-k8s-1 to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:11:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: 2023-01-05T09:11:39Z [verbose] Add: openshift-monitoring:alertmanager-main-1:b59f4808-02df-4b2e-9694-8cf72a1b5e75:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/a22d0db9-a46e-42fb-931a-4c5793845fd1"}],"ips":[{"version":"4","interface":0,"address":"10.129.2.13/23","gateway":"10.129.2.1"}],"dns":{}} Jan 05 09:11:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: I0105 09:11:39.219320 71604 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"alertmanager-main-1", UID:"b59f4808-02df-4b2e-9694-8cf72a1b5e75", APIVersion:"v1", ResourceVersion:"73844", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.129.2.13/23] from openshift-sdn Jan 05 09:11:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:39.237270591Z" level=info msg="Got pod network &{Name:alertmanager-main-1 Namespace:openshift-monitoring ID:4cc6b8c0009aed4f5500066da7f716df203dd1969281bb8dbce4d7052e55ec50 UID:b59f4808-02df-4b2e-9694-8cf72a1b5e75 NetNS:/var/run/netns/a22d0db9-a46e-42fb-931a-4c5793845fd1 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:11:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:39.237441536Z" level=info msg="Checking pod openshift-monitoring_alertmanager-main-1 for CNI network multus-cni-network (type=multus)" Jan 05 09:11:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:39.237652 1425 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/alertmanager-main-1] Jan 05 09:11:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:11:39.241471 1425 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb59f4808_02df_4b2e_9694_8cf72a1b5e75.slice/crio-4cc6b8c0009aed4f5500066da7f716df203dd1969281bb8dbce4d7052e55ec50.scope WatchSource:0}: Error finding container 4cc6b8c0009aed4f5500066da7f716df203dd1969281bb8dbce4d7052e55ec50: Status 404 returned error can't find the container with id 4cc6b8c0009aed4f5500066da7f716df203dd1969281bb8dbce4d7052e55ec50 Jan 05 09:11:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:39.245311141Z" level=info msg="Ran pod sandbox 4cc6b8c0009aed4f5500066da7f716df203dd1969281bb8dbce4d7052e55ec50 with infra container: openshift-monitoring/alertmanager-main-1/POD" id=32bd009c-ff02-4645-af29-df6fcc619622 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:11:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:39.246213199Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e6173b5d77def9e8058c7ff82f1ad65f537d1a81f88fcdabe6d8e30bff49ca89" id=b5012200-cd3c-4411-9afc-7780ae5babb0 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:11:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:39.246395034Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e6173b5d77def9e8058c7ff82f1ad65f537d1a81f88fcdabe6d8e30bff49ca89 not found" id=b5012200-cd3c-4411-9afc-7780ae5babb0 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:11:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:39.247013190Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e6173b5d77def9e8058c7ff82f1ad65f537d1a81f88fcdabe6d8e30bff49ca89" id=cb6d0ab9-5a39-4625-b03a-869d3a70dfc7 name=/runtime.v1.ImageService/PullImage Jan 05 09:11:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:39.247929526Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e6173b5d77def9e8058c7ff82f1ad65f537d1a81f88fcdabe6d8e30bff49ca89\"" Jan 05 09:11:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:39.652748793Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e6173b5d77def9e8058c7ff82f1ad65f537d1a81f88fcdabe6d8e30bff49ca89\"" Jan 05 09:11:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:39.813601 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-1" event=&{ID:b59f4808-02df-4b2e-9694-8cf72a1b5e75 Type:ContainerStarted Data:4cc6b8c0009aed4f5500066da7f716df203dd1969281bb8dbce4d7052e55ec50} Jan 05 09:11:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: 2023-01-05T09:11:40Z [verbose] Add: openshift-monitoring:prometheus-k8s-1:8ac551e8-a959-4212-86a1-2b3ac482586f:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/cf32a4f0-a7ea-4d0a-ada0-ddff6598e5f3"}],"ips":[{"version":"4","interface":0,"address":"10.129.2.14/23","gateway":"10.129.2.1"}],"dns":{}} Jan 05 09:11:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: I0105 09:11:40.800215 71680 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"prometheus-k8s-1", UID:"8ac551e8-a959-4212-86a1-2b3ac482586f", APIVersion:"v1", ResourceVersion:"73893", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.129.2.14/23] from openshift-sdn Jan 05 09:11:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:40.823539112Z" level=info msg="Got pod network &{Name:prometheus-k8s-1 Namespace:openshift-monitoring ID:bf998995ead8afb2642821b82636d7b867d8c0aa5ab164fe09e7396c2701f63b UID:8ac551e8-a959-4212-86a1-2b3ac482586f NetNS:/var/run/netns/cf32a4f0-a7ea-4d0a-ada0-ddff6598e5f3 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:11:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:40.823713651Z" level=info msg="Checking pod openshift-monitoring_prometheus-k8s-1 for CNI network multus-cni-network (type=multus)" Jan 05 09:11:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:40.836396 1425 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/prometheus-k8s-1] Jan 05 09:11:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:40.838003981Z" level=info msg="Ran pod sandbox bf998995ead8afb2642821b82636d7b867d8c0aa5ab164fe09e7396c2701f63b with infra container: openshift-monitoring/prometheus-k8s-1/POD" id=326dcfe0-3314-4230-b163-08231b9e8df1 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:11:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:40.847081417Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=587033ec-8e9e-4a54-aa48-2ab1e2cd6cd3 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:11:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:40.847279037Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f not found" id=587033ec-8e9e-4a54-aa48-2ab1e2cd6cd3 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:11:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:40.848645579Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=aaee8c7d-75b9-4420-a1ef-3fbd5495c0f3 name=/runtime.v1.ImageService/PullImage Jan 05 09:11:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:40.849688017Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f\"" Jan 05 09:11:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:41.198513030Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f\"" Jan 05 09:11:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:41.820471 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-1" event=&{ID:8ac551e8-a959-4212-86a1-2b3ac482586f Type:ContainerStarted Data:bf998995ead8afb2642821b82636d7b867d8c0aa5ab164fe09e7396c2701f63b} Jan 05 09:11:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:42.842957941Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d0336c488437103651d61c3d74b49c13559c7cbc866106918883771b271d0ae9" id=bd4fc555-7cce-4b5c-b572-dd5a7e32f4f3 name=/runtime.v1.ImageService/PullImage Jan 05 09:11:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:42.843807318Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d0336c488437103651d61c3d74b49c13559c7cbc866106918883771b271d0ae9" id=ee004506-9690-4931-8b99-d3eb04fd6878 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:11:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:42.845084633Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:65e56f749865ac86fccf0a7cd0a7e5eb6a9964cef9de123547513ceabc7c7051,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d0336c488437103651d61c3d74b49c13559c7cbc866106918883771b271d0ae9],Size_:314684149,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=ee004506-9690-4931-8b99-d3eb04fd6878 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:11:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:42.845879561Z" level=info msg="Creating container: openshift-kube-storage-version-migrator/migrator-6cd87bdff8-p9dkf/migrator" id=68d9239b-8ae7-4c37-8c45-5c14bbab2f33 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:11:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:42.845989618Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:11:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:43.027069582Z" level=info msg="Created container ff17947445f945fe096ba0ee1a903135c1548bf46b79e37353e7eddd5042f6f8: openshift-kube-storage-version-migrator/migrator-6cd87bdff8-p9dkf/migrator" id=68d9239b-8ae7-4c37-8c45-5c14bbab2f33 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:11:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:43.027871765Z" level=info msg="Starting container: ff17947445f945fe096ba0ee1a903135c1548bf46b79e37353e7eddd5042f6f8" id=0653e187-bf14-4268-8211-d51275977c2e name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:11:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:43.040708828Z" level=info msg="Started container" PID=71826 containerID=ff17947445f945fe096ba0ee1a903135c1548bf46b79e37353e7eddd5042f6f8 description=openshift-kube-storage-version-migrator/migrator-6cd87bdff8-p9dkf/migrator id=0653e187-bf14-4268-8211-d51275977c2e name=/runtime.v1.RuntimeService/StartContainer sandboxID=148c6bc4da65fd9b07c5fca9d24c7096ee2b924f8d407958ac234e9758e340f4 Jan 05 09:11:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:43.826855 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-6cd87bdff8-p9dkf" event=&{ID:26f1eb05-de00-4fc2-864c-6b97e7f82e96 Type:ContainerStarted Data:ff17947445f945fe096ba0ee1a903135c1548bf46b79e37353e7eddd5042f6f8} Jan 05 09:11:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:44.845196 1425 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-multus/cni-sysctl-allowlist-ds-nt5pz] Jan 05 09:11:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:44.845242 1425 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:11:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:44.963232 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frtl7\" (UniqueName: \"kubernetes.io/projected/a3cb294f-6846-4e6b-b42f-434294a1cefd-kube-api-access-frtl7\") pod \"cni-sysctl-allowlist-ds-nt5pz\" (UID: \"a3cb294f-6846-4e6b-b42f-434294a1cefd\") " pod="openshift-multus/cni-sysctl-allowlist-ds-nt5pz" Jan 05 09:11:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:44.963302 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/a3cb294f-6846-4e6b-b42f-434294a1cefd-ready\") pod \"cni-sysctl-allowlist-ds-nt5pz\" (UID: \"a3cb294f-6846-4e6b-b42f-434294a1cefd\") " pod="openshift-multus/cni-sysctl-allowlist-ds-nt5pz" Jan 05 09:11:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:44.963414 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a3cb294f-6846-4e6b-b42f-434294a1cefd-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-nt5pz\" (UID: \"a3cb294f-6846-4e6b-b42f-434294a1cefd\") " pod="openshift-multus/cni-sysctl-allowlist-ds-nt5pz" Jan 05 09:11:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:44.963478 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a3cb294f-6846-4e6b-b42f-434294a1cefd-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-nt5pz\" (UID: \"a3cb294f-6846-4e6b-b42f-434294a1cefd\") " pod="openshift-multus/cni-sysctl-allowlist-ds-nt5pz" Jan 05 09:11:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:45.064727 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a3cb294f-6846-4e6b-b42f-434294a1cefd-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-nt5pz\" (UID: \"a3cb294f-6846-4e6b-b42f-434294a1cefd\") " pod="openshift-multus/cni-sysctl-allowlist-ds-nt5pz" Jan 05 09:11:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:45.064788 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a3cb294f-6846-4e6b-b42f-434294a1cefd-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-nt5pz\" (UID: \"a3cb294f-6846-4e6b-b42f-434294a1cefd\") " pod="openshift-multus/cni-sysctl-allowlist-ds-nt5pz" Jan 05 09:11:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:45.064841 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-frtl7\" (UniqueName: \"kubernetes.io/projected/a3cb294f-6846-4e6b-b42f-434294a1cefd-kube-api-access-frtl7\") pod \"cni-sysctl-allowlist-ds-nt5pz\" (UID: \"a3cb294f-6846-4e6b-b42f-434294a1cefd\") " pod="openshift-multus/cni-sysctl-allowlist-ds-nt5pz" Jan 05 09:11:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:45.064886 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/a3cb294f-6846-4e6b-b42f-434294a1cefd-ready\") pod \"cni-sysctl-allowlist-ds-nt5pz\" (UID: \"a3cb294f-6846-4e6b-b42f-434294a1cefd\") " pod="openshift-multus/cni-sysctl-allowlist-ds-nt5pz" Jan 05 09:11:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:45.065218 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/a3cb294f-6846-4e6b-b42f-434294a1cefd-ready\") pod \"cni-sysctl-allowlist-ds-nt5pz\" (UID: \"a3cb294f-6846-4e6b-b42f-434294a1cefd\") " pod="openshift-multus/cni-sysctl-allowlist-ds-nt5pz" Jan 05 09:11:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:45.065713 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a3cb294f-6846-4e6b-b42f-434294a1cefd-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-nt5pz\" (UID: \"a3cb294f-6846-4e6b-b42f-434294a1cefd\") " pod="openshift-multus/cni-sysctl-allowlist-ds-nt5pz" Jan 05 09:11:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:45.065778 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a3cb294f-6846-4e6b-b42f-434294a1cefd-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-nt5pz\" (UID: \"a3cb294f-6846-4e6b-b42f-434294a1cefd\") " pod="openshift-multus/cni-sysctl-allowlist-ds-nt5pz" Jan 05 09:11:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:45.086707 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-frtl7\" (UniqueName: \"kubernetes.io/projected/a3cb294f-6846-4e6b-b42f-434294a1cefd-kube-api-access-frtl7\") pod \"cni-sysctl-allowlist-ds-nt5pz\" (UID: \"a3cb294f-6846-4e6b-b42f-434294a1cefd\") " pod="openshift-multus/cni-sysctl-allowlist-ds-nt5pz" Jan 05 09:11:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:45.172780 1425 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-nt5pz" Jan 05 09:11:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:45.173450689Z" level=info msg="Running pod sandbox: openshift-multus/cni-sysctl-allowlist-ds-nt5pz/POD" id=c06e9922-5c02-4b89-bd68-3238ae2defbc name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:11:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:45.173513733Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:11:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:45.199037423Z" level=info msg="Got pod network &{Name:cni-sysctl-allowlist-ds-nt5pz Namespace:openshift-multus ID:52159b4d1e3648c7df20e17bffb14777d3a3c663e803b06565beb65f460be913 UID:a3cb294f-6846-4e6b-b42f-434294a1cefd NetNS:/var/run/netns/6ef4c066-c3e6-40db-a7e4-24e61dbd531e Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:11:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:45.199081817Z" level=info msg="Adding pod openshift-multus_cni-sysctl-allowlist-ds-nt5pz to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:11:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:46.856519 1425 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-multus/cni-sysctl-allowlist-ds-nt5pz] Jan 05 09:11:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: 2023-01-05T09:11:46Z [verbose] Add: openshift-multus:cni-sysctl-allowlist-ds-nt5pz:a3cb294f-6846-4e6b-b42f-434294a1cefd:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/6ef4c066-c3e6-40db-a7e4-24e61dbd531e"}],"ips":[{"version":"4","interface":0,"address":"10.129.2.15/23","gateway":"10.129.2.1"}],"dns":{}} Jan 05 09:11:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: I0105 09:11:46.839665 71867 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-multus", Name:"cni-sysctl-allowlist-ds-nt5pz", UID:"a3cb294f-6846-4e6b-b42f-434294a1cefd", APIVersion:"v1", ResourceVersion:"74044", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.129.2.15/23] from openshift-sdn Jan 05 09:11:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:46.857253746Z" level=info msg="Got pod network &{Name:cni-sysctl-allowlist-ds-nt5pz Namespace:openshift-multus ID:52159b4d1e3648c7df20e17bffb14777d3a3c663e803b06565beb65f460be913 UID:a3cb294f-6846-4e6b-b42f-434294a1cefd NetNS:/var/run/netns/6ef4c066-c3e6-40db-a7e4-24e61dbd531e Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:11:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:46.857407205Z" level=info msg="Checking pod openshift-multus_cni-sysctl-allowlist-ds-nt5pz for CNI network multus-cni-network (type=multus)" Jan 05 09:11:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:46.865592797Z" level=info msg="Ran pod sandbox 52159b4d1e3648c7df20e17bffb14777d3a3c663e803b06565beb65f460be913 with infra container: openshift-multus/cni-sysctl-allowlist-ds-nt5pz/POD" id=c06e9922-5c02-4b89-bd68-3238ae2defbc name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:11:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:46.869975564Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=b17127e6-f055-4a82-b9ec-1c61aac97ede name=/runtime.v1.ImageService/ImageStatus Jan 05 09:11:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:46.870190984Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b17127e6-f055-4a82-b9ec-1c61aac97ede name=/runtime.v1.ImageService/ImageStatus Jan 05 09:11:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:46.870949518Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=11415edd-e171-4e78-9d9e-54b56c3c1177 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:11:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:46.871106328Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=11415edd-e171-4e78-9d9e-54b56c3c1177 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:11:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:46.871937314Z" level=info msg="Creating container: openshift-multus/cni-sysctl-allowlist-ds-nt5pz/kube-multus-additional-cni-plugins" id=a4899b37-7863-4e59-b830-d1049c05ff74 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:11:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:46.872065901Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:11:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:47.077779435Z" level=info msg="Created container 12ba0576c39c7251a305990e9512ca687002ea4e088fb3b78d3d7f7b77efabc5: openshift-multus/cni-sysctl-allowlist-ds-nt5pz/kube-multus-additional-cni-plugins" id=a4899b37-7863-4e59-b830-d1049c05ff74 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:11:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:47.078253972Z" level=info msg="Starting container: 12ba0576c39c7251a305990e9512ca687002ea4e088fb3b78d3d7f7b77efabc5" id=13454f26-f9f8-4fe3-ad06-0b5a715cca37 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:11:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:47.088826250Z" level=info msg="Started container" PID=71947 containerID=12ba0576c39c7251a305990e9512ca687002ea4e088fb3b78d3d7f7b77efabc5 description=openshift-multus/cni-sysctl-allowlist-ds-nt5pz/kube-multus-additional-cni-plugins id=13454f26-f9f8-4fe3-ad06-0b5a715cca37 name=/runtime.v1.RuntimeService/StartContainer sandboxID=52159b4d1e3648c7df20e17bffb14777d3a3c663e803b06565beb65f460be913 Jan 05 09:11:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:47.838054 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-nt5pz" event=&{ID:a3cb294f-6846-4e6b-b42f-434294a1cefd Type:ContainerStarted Data:12ba0576c39c7251a305990e9512ca687002ea4e088fb3b78d3d7f7b77efabc5} Jan 05 09:11:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:47.838089 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-nt5pz" event=&{ID:a3cb294f-6846-4e6b-b42f-434294a1cefd Type:ContainerStarted Data:52159b4d1e3648c7df20e17bffb14777d3a3c663e803b06565beb65f460be913} Jan 05 09:11:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:47.838323 1425 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-multus/cni-sysctl-allowlist-ds-nt5pz" Jan 05 09:11:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:48.883800 1425 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-multus/cni-sysctl-allowlist-ds-nt5pz" Jan 05 09:11:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:49.886416 1425 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-multus/cni-sysctl-allowlist-ds-nt5pz] Jan 05 09:11:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:50.843899 1425 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-multus/cni-sysctl-allowlist-ds-nt5pz" podUID=a3cb294f-6846-4e6b-b42f-434294a1cefd containerName="kube-multus-additional-cni-plugins" containerID="cri-o://12ba0576c39c7251a305990e9512ca687002ea4e088fb3b78d3d7f7b77efabc5" gracePeriod=30 Jan 05 09:11:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:50.844470952Z" level=info msg="Stopping container: 12ba0576c39c7251a305990e9512ca687002ea4e088fb3b78d3d7f7b77efabc5 (timeout: 30s)" id=7b6c9e91-9144-4f2c-8e5e-a06968c8662a name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:11:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:11:54.481133 1425 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-64655f6865-dx2q6" Jan 05 09:11:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:57.634159665Z" level=warning msg="Found defunct process with PID 72291 (iptables)" Jan 05 09:11:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:59.556967159Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9696f8f231d49aadf63ab980ec5c17c4fd87011930daecd30cc9e4ee102f1aeb" id=9fb1448d-d561-4c4e-8bed-b69ea7bc0ee6 name=/runtime.v1.ImageService/PullImage Jan 05 09:11:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:59.558278806Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9696f8f231d49aadf63ab980ec5c17c4fd87011930daecd30cc9e4ee102f1aeb" id=c04842c3-30bf-4ac2-b6ab-d3cebb647b1f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:11:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:59.560600441Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:ea0b364e3a8a4e4b3f4ab3288e42c5be8021187d32791029f9cb9a500c8da100,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9696f8f231d49aadf63ab980ec5c17c4fd87011930daecd30cc9e4ee102f1aeb],Size_:402351209,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=c04842c3-30bf-4ac2-b6ab-d3cebb647b1f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:11:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:59.562565938Z" level=info msg="Creating container: openshift-monitoring/prometheus-adapter-5894cc86c7-47422/prometheus-adapter" id=0fb0b833-836d-4f6c-a431-14ed7f2c2747 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:11:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:59.562692719Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:11:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:59.846717044Z" level=info msg="Created container a30b14e5cf7f780de074100612e1db98e849c8b763e7b551ab0f6558754f69f6: openshift-monitoring/prometheus-adapter-5894cc86c7-47422/prometheus-adapter" id=0fb0b833-836d-4f6c-a431-14ed7f2c2747 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:11:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:59.853283530Z" level=info msg="Starting container: a30b14e5cf7f780de074100612e1db98e849c8b763e7b551ab0f6558754f69f6" id=42edf764-3274-4e44-945f-dda10e495376 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:11:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:11:59.948933049Z" level=info msg="Started container" PID=72360 containerID=a30b14e5cf7f780de074100612e1db98e849c8b763e7b551ab0f6558754f69f6 description=openshift-monitoring/prometheus-adapter-5894cc86c7-47422/prometheus-adapter id=42edf764-3274-4e44-945f-dda10e495376 name=/runtime.v1.RuntimeService/StartContainer sandboxID=ef7044b787198dfb2aab7533e9c72c07b8c2ef842ecb652a788dc647c485157c Jan 05 09:12:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:01.032394 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" event=&{ID:a82fd83b-5344-4e7a-91dd-5894f3846e87 Type:ContainerStarted Data:a30b14e5cf7f780de074100612e1db98e849c8b763e7b551ab0f6558754f69f6} Jan 05 09:12:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:01.033792 1425 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" Jan 05 09:12:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:02.892154375Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:bfd14188641a2d34e5b2b2ed3336b3b440155533f3fd7c6de776300dc3ff9e3b" id=477433b0-3246-4ca0-adf1-eee7b7dbd563 name=/runtime.v1.ImageService/PullImage Jan 05 09:12:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:02.893090471Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:bfd14188641a2d34e5b2b2ed3336b3b440155533f3fd7c6de776300dc3ff9e3b" id=88beb64a-1ae8-4090-b14e-b3c8d26465bf name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:02.895312045Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:ad00ebc82f5cc671785b8401477b42db226d1c335ea60a4bb39bf28a961de840,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:bfd14188641a2d34e5b2b2ed3336b3b440155533f3fd7c6de776300dc3ff9e3b],Size_:411774965,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=88beb64a-1ae8-4090-b14e-b3c8d26465bf name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:02.896329069Z" level=info msg="Creating container: openshift-ingress/router-default-5cf6bbdff9-fkmcn/router" id=ff4901ca-cc7e-4b54-bcec-aa37fee0164b name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:02.896457495Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:12:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:03.008901937Z" level=info msg="Created container 8c886b3e9fbca71166a4d01928168d286331774947ddef0326c331f3624f5f6a: openshift-ingress/router-default-5cf6bbdff9-fkmcn/router" id=ff4901ca-cc7e-4b54-bcec-aa37fee0164b name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:03.009495282Z" level=info msg="Starting container: 8c886b3e9fbca71166a4d01928168d286331774947ddef0326c331f3624f5f6a" id=44780cba-ff56-4b40-ac6f-8f7b26b4a60c name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:12:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:03.019834625Z" level=info msg="Started container" PID=72586 containerID=8c886b3e9fbca71166a4d01928168d286331774947ddef0326c331f3624f5f6a description=openshift-ingress/router-default-5cf6bbdff9-fkmcn/router id=44780cba-ff56-4b40-ac6f-8f7b26b4a60c name=/runtime.v1.RuntimeService/StartContainer sandboxID=0412a26df3b1c530016b6935b36a365c8ab277d3b739d57e9cec85e5975d314c Jan 05 09:12:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:03.049226 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" event=&{ID:66d22b70-08bb-4199-b258-ac4590e9b487 Type:ContainerStarted Data:8c886b3e9fbca71166a4d01928168d286331774947ddef0326c331f3624f5f6a} Jan 05 09:12:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:03.759837 1425 kubelet.go:2229] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" Jan 05 09:12:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:03.777154 1425 kubelet.go:2229] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" Jan 05 09:12:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:04.073591 1425 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" Jan 05 09:12:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:04.076438 1425 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" Jan 05 09:12:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:05.579836437Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:dbf46164dcc3b2b6ef5df5d043ddc4fd859fd2c73c4862808424b3df6a1e27e7" id=67f26b37-1aa8-4f5e-83fc-9922fe8fb8b8 name=/runtime.v1.ImageService/PullImage Jan 05 09:12:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:05.581333262Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e6173b5d77def9e8058c7ff82f1ad65f537d1a81f88fcdabe6d8e30bff49ca89" id=cb6d0ab9-5a39-4625-b03a-869d3a70dfc7 name=/runtime.v1.ImageService/PullImage Jan 05 09:12:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:05.584791331Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=aaee8c7d-75b9-4420-a1ef-3fbd5495c0f3 name=/runtime.v1.ImageService/PullImage Jan 05 09:12:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:05.585989028Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=d25cf69e-805a-47d6-be26-1dcd1eca7c48 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:05.587676928Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=d25cf69e-805a-47d6-be26-1dcd1eca7c48 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:05.587748993Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:dbf46164dcc3b2b6ef5df5d043ddc4fd859fd2c73c4862808424b3df6a1e27e7" id=71acd167-cb0d-45a0-99c4-2678226630a4 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:05.589464587Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:20e9403b34a0d9d25151a4ddbee3f69b99894fd8e058e6cb7310fe95080ecb03,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:dbf46164dcc3b2b6ef5df5d043ddc4fd859fd2c73c4862808424b3df6a1e27e7],Size_:374119253,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=71acd167-cb0d-45a0-99c4-2678226630a4 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:05.589537315Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-1/init-config-reloader" id=4610c1b2-e954-4c69-9eef-dc12f31bd748 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:05.589662736Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:12:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:05.601439106Z" level=info msg="Creating container: openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-54rjq/prometheus-operator-admission-webhook" id=31180b07-7a95-46fc-a20d-95249cf06bed name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:05.601555126Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:12:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:05.612733154Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e6173b5d77def9e8058c7ff82f1ad65f537d1a81f88fcdabe6d8e30bff49ca89" id=67876f9b-eedb-4da0-9df3-bd871c7df768 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:05.628999302Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:d168c738461c25cc0c3fa96cfcdf2dc337ca08f5bf3a6b550fcb29a59ea39270,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e6173b5d77def9e8058c7ff82f1ad65f537d1a81f88fcdabe6d8e30bff49ca89],Size_:333780768,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=67876f9b-eedb-4da0-9df3-bd871c7df768 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:05.631330240Z" level=info msg="Creating container: openshift-monitoring/alertmanager-main-1/alertmanager" id=8b1ec08b-c9a9-41c8-bd2b-392829c081f8 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:05.631450030Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:12:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:05.830299597Z" level=info msg="Created container c2fc2489b690ab18d10fc72750d26008cb07a65be8420998a29168a6a7cea2af: openshift-monitoring/prometheus-k8s-1/init-config-reloader" id=4610c1b2-e954-4c69-9eef-dc12f31bd748 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:05.830631184Z" level=info msg="Created container 902fedd514d362305b82903fec325b5adb139bd729665753e388c28b1e4d58bf: openshift-monitoring/alertmanager-main-1/alertmanager" id=8b1ec08b-c9a9-41c8-bd2b-392829c081f8 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:05.830942152Z" level=info msg="Starting container: c2fc2489b690ab18d10fc72750d26008cb07a65be8420998a29168a6a7cea2af" id=ef19ad6a-9fe8-4ea2-aaed-10db15e2c01c name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:12:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:05.831222437Z" level=info msg="Starting container: 902fedd514d362305b82903fec325b5adb139bd729665753e388c28b1e4d58bf" id=e238beab-b0ca-4522-ad62-13abbcb4e906 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:12:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:05.850513821Z" level=info msg="Started container" PID=72790 containerID=c2fc2489b690ab18d10fc72750d26008cb07a65be8420998a29168a6a7cea2af description=openshift-monitoring/prometheus-k8s-1/init-config-reloader id=ef19ad6a-9fe8-4ea2-aaed-10db15e2c01c name=/runtime.v1.RuntimeService/StartContainer sandboxID=bf998995ead8afb2642821b82636d7b867d8c0aa5ab164fe09e7396c2701f63b Jan 05 09:12:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:05.881764925Z" level=info msg="Started container" PID=72784 containerID=902fedd514d362305b82903fec325b5adb139bd729665753e388c28b1e4d58bf description=openshift-monitoring/alertmanager-main-1/alertmanager id=e238beab-b0ca-4522-ad62-13abbcb4e906 name=/runtime.v1.RuntimeService/StartContainer sandboxID=4cc6b8c0009aed4f5500066da7f716df203dd1969281bb8dbce4d7052e55ec50 Jan 05 09:12:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:05.896445516Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=91563152-0fc5-4ca7-b486-419041ba8534 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:05.898734584Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=91563152-0fc5-4ca7-b486-419041ba8534 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:05.907201839Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=5ebdd38f-c5b8-45d3-843d-fc10f2b8c05a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:05.909519650Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=5ebdd38f-c5b8-45d3-843d-fc10f2b8c05a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:05.910632471Z" level=info msg="Created container b6613cda35aa605c0c2f9fa03950bbd6def356ba0a1a95a665f8c80d6b17a14f: openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-54rjq/prometheus-operator-admission-webhook" id=31180b07-7a95-46fc-a20d-95249cf06bed name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:05.910880378Z" level=info msg="Creating container: openshift-monitoring/alertmanager-main-1/config-reloader" id=1c733756-1485-4b13-b643-f14d5163a59e name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:05.910986292Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:12:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:05.915838835Z" level=info msg="Starting container: b6613cda35aa605c0c2f9fa03950bbd6def356ba0a1a95a665f8c80d6b17a14f" id=ec8657b7-46b3-4c47-b772-b0a98fa9f5e7 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:12:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:05.939809287Z" level=info msg="Started container" PID=72792 containerID=b6613cda35aa605c0c2f9fa03950bbd6def356ba0a1a95a665f8c80d6b17a14f description=openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-54rjq/prometheus-operator-admission-webhook id=ec8657b7-46b3-4c47-b772-b0a98fa9f5e7 name=/runtime.v1.RuntimeService/StartContainer sandboxID=8e21977f4f2aa77ecd7154452d4a15adc91d3b323c169dda297ad9c1f8901298 Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.074214098Z" level=info msg="Created container e9efd1bb280b0736df20aa595c1c6f65c0d953e8a7a6a1dde105be0f728d6de6: openshift-monitoring/alertmanager-main-1/config-reloader" id=1c733756-1485-4b13-b643-f14d5163a59e name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.074810251Z" level=info msg="Starting container: e9efd1bb280b0736df20aa595c1c6f65c0d953e8a7a6a1dde105be0f728d6de6" id=89ec6125-e79a-4170-909d-4a8bf19f891c name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:06.081281 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-54rjq" event=&{ID:f8e5be3f-9ab7-484c-b61b-6231a8774e24 Type:ContainerStarted Data:b6613cda35aa605c0c2f9fa03950bbd6def356ba0a1a95a665f8c80d6b17a14f} Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:06.082587 1425 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-54rjq" Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:06.084068 1425 generic.go:296] "Generic (PLEG): container finished" podID=8ac551e8-a959-4212-86a1-2b3ac482586f containerID="c2fc2489b690ab18d10fc72750d26008cb07a65be8420998a29168a6a7cea2af" exitCode=0 Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:06.084125 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-1" event=&{ID:8ac551e8-a959-4212-86a1-2b3ac482586f Type:ContainerDied Data:c2fc2489b690ab18d10fc72750d26008cb07a65be8420998a29168a6a7cea2af} Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:06.091657 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-1" event=&{ID:b59f4808-02df-4b2e-9694-8cf72a1b5e75 Type:ContainerStarted Data:902fedd514d362305b82903fec325b5adb139bd729665753e388c28b1e4d58bf} Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:06.094313 1425 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-54rjq" Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.084889456Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:91e7d68b28aa00e209065604b2b9ac1400671d9102581e8b49b37726acd7b6e4" id=74468d55-c7be-4114-bf49-b7226c6e9b99 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.085105347Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:91e7d68b28aa00e209065604b2b9ac1400671d9102581e8b49b37726acd7b6e4 not found" id=74468d55-c7be-4114-bf49-b7226c6e9b99 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.087753295Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:91e7d68b28aa00e209065604b2b9ac1400671d9102581e8b49b37726acd7b6e4" id=976302a0-fd40-442e-bd3c-20729ecc0d1d name=/runtime.v1.ImageService/PullImage Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.089114462Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:91e7d68b28aa00e209065604b2b9ac1400671d9102581e8b49b37726acd7b6e4\"" Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.097806309Z" level=info msg="Started container" PID=72883 containerID=e9efd1bb280b0736df20aa595c1c6f65c0d953e8a7a6a1dde105be0f728d6de6 description=openshift-monitoring/alertmanager-main-1/config-reloader id=89ec6125-e79a-4170-909d-4a8bf19f891c name=/runtime.v1.RuntimeService/StartContainer sandboxID=4cc6b8c0009aed4f5500066da7f716df203dd1969281bb8dbce4d7052e55ec50 Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.113607511Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=81c4f8e7-eba7-4963-8a97-2c10a31e42ec name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.113827403Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=81c4f8e7-eba7-4963-8a97-2c10a31e42ec name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.114768395Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=436b15b4-d0b4-48e3-b63e-b03d5caf77b3 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.114965547Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=436b15b4-d0b4-48e3-b63e-b03d5caf77b3 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.117015043Z" level=info msg="Creating container: openshift-monitoring/alertmanager-main-1/alertmanager-proxy" id=3f024a8b-feb8-40cd-a69c-21e0279c4637 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.117152453Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.255222050Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:91e7d68b28aa00e209065604b2b9ac1400671d9102581e8b49b37726acd7b6e4\"" Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.351877435Z" level=info msg="Created container af5d20453c1aa7ea8d99d15e82767c98893d319f38bb9cd97aa815202e28d32c: openshift-monitoring/alertmanager-main-1/alertmanager-proxy" id=3f024a8b-feb8-40cd-a69c-21e0279c4637 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.352311730Z" level=info msg="Starting container: af5d20453c1aa7ea8d99d15e82767c98893d319f38bb9cd97aa815202e28d32c" id=c81297b8-f76d-4202-b81f-8990b69a8299 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.361500690Z" level=info msg="Started container" PID=72945 containerID=af5d20453c1aa7ea8d99d15e82767c98893d319f38bb9cd97aa815202e28d32c description=openshift-monitoring/alertmanager-main-1/alertmanager-proxy id=c81297b8-f76d-4202-b81f-8990b69a8299 name=/runtime.v1.RuntimeService/StartContainer sandboxID=4cc6b8c0009aed4f5500066da7f716df203dd1969281bb8dbce4d7052e55ec50 Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.386776133Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=1a3f9cad-1234-4b73-929f-1466ae433cc3 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.387027929Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=1a3f9cad-1234-4b73-929f-1466ae433cc3 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.388208435Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=28c766e4-1b33-47a9-80d5-cbe6ffafa9dc name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.388430377Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=28c766e4-1b33-47a9-80d5-cbe6ffafa9dc name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.389594315Z" level=info msg="Creating container: openshift-monitoring/alertmanager-main-1/kube-rbac-proxy" id=369c837e-1196-4f8e-831d-ce07fc0d188f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.389708243Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.510240426Z" level=info msg="Created container 6eba543bd270671b29706b350f4dfdf034230e4d33c811838d47249838c4ff2b: openshift-monitoring/alertmanager-main-1/kube-rbac-proxy" id=369c837e-1196-4f8e-831d-ce07fc0d188f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.510688179Z" level=info msg="Starting container: 6eba543bd270671b29706b350f4dfdf034230e4d33c811838d47249838c4ff2b" id=030965e3-78f6-46bd-997d-0d10dd1b4fe9 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.519764897Z" level=info msg="Started container" PID=72989 containerID=6eba543bd270671b29706b350f4dfdf034230e4d33c811838d47249838c4ff2b description=openshift-monitoring/alertmanager-main-1/kube-rbac-proxy id=030965e3-78f6-46bd-997d-0d10dd1b4fe9 name=/runtime.v1.RuntimeService/StartContainer sandboxID=4cc6b8c0009aed4f5500066da7f716df203dd1969281bb8dbce4d7052e55ec50 Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.544182519Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=401cf9de-93bc-49bf-8e59-7acb9053c233 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.544449745Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=401cf9de-93bc-49bf-8e59-7acb9053c233 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.545327513Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=ff624c38-0777-440a-b5cc-5d0ac5662d49 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.545535244Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=ff624c38-0777-440a-b5cc-5d0ac5662d49 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.546917505Z" level=info msg="Creating container: openshift-monitoring/alertmanager-main-1/kube-rbac-proxy-metric" id=58f48e8c-53c6-4360-8aec-45794909517a name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.547026274Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.661638481Z" level=info msg="Created container 064809cddb0902a863e4fcd8d1da29b6a436aa3118d40b3b5ab6fbbda4fbf3d9: openshift-monitoring/alertmanager-main-1/kube-rbac-proxy-metric" id=58f48e8c-53c6-4360-8aec-45794909517a name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.662257064Z" level=info msg="Starting container: 064809cddb0902a863e4fcd8d1da29b6a436aa3118d40b3b5ab6fbbda4fbf3d9" id=7c9bb4c1-cf0a-4866-94d5-2e8cd0a56198 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.674213971Z" level=info msg="Started container" PID=73027 containerID=064809cddb0902a863e4fcd8d1da29b6a436aa3118d40b3b5ab6fbbda4fbf3d9 description=openshift-monitoring/alertmanager-main-1/kube-rbac-proxy-metric id=7c9bb4c1-cf0a-4866-94d5-2e8cd0a56198 name=/runtime.v1.RuntimeService/StartContainer sandboxID=4cc6b8c0009aed4f5500066da7f716df203dd1969281bb8dbce4d7052e55ec50 Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.695328130Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816" id=7e12297b-f94e-4356-b61a-2e3a8f3f19d1 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.695591089Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816 not found" id=7e12297b-f94e-4356-b61a-2e3a8f3f19d1 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.696612171Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816" id=c25a47f6-ce78-4ca8-b4f6-7405b90e665d name=/runtime.v1.ImageService/PullImage Jan 05 09:12:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:06.697612509Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816\"" Jan 05 09:12:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:07.095822 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-1" event=&{ID:b59f4808-02df-4b2e-9694-8cf72a1b5e75 Type:ContainerStarted Data:e9efd1bb280b0736df20aa595c1c6f65c0d953e8a7a6a1dde105be0f728d6de6} Jan 05 09:12:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:07.095858 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-1" event=&{ID:b59f4808-02df-4b2e-9694-8cf72a1b5e75 Type:ContainerStarted Data:064809cddb0902a863e4fcd8d1da29b6a436aa3118d40b3b5ab6fbbda4fbf3d9} Jan 05 09:12:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:07.095876 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-1" event=&{ID:b59f4808-02df-4b2e-9694-8cf72a1b5e75 Type:ContainerStarted Data:6eba543bd270671b29706b350f4dfdf034230e4d33c811838d47249838c4ff2b} Jan 05 09:12:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:07.095892 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-1" event=&{ID:b59f4808-02df-4b2e-9694-8cf72a1b5e75 Type:ContainerStarted Data:af5d20453c1aa7ea8d99d15e82767c98893d319f38bb9cd97aa815202e28d32c} Jan 05 09:12:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:07.913433476Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816\"" Jan 05 09:12:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:10.175789381Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc" id=c5819c43-72d5-457a-92f3-f99100961920 name=/runtime.v1.ImageService/PullImage Jan 05 09:12:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:10.176895183Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc" id=02313da6-e5d6-485f-9c17-cf869c465f90 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:10.178665863Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:08ff7a43d9185529520a29e114f0c28b070000b464a44b347240015f3b466488,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc],Size_:368278089,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=02313da6-e5d6-485f-9c17-cf869c465f90 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:10.179943233Z" level=info msg="Creating container: openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/thanos-query" id=2bb65738-15dc-4488-ae8d-60a9fc3d3ea5 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:10.180038454Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:12:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:10.397778548Z" level=info msg="Created container fcb76a52605ae84e7415c0b6e56aa8e1f010607edc042a479cb6e4adde24ecc4: openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/thanos-query" id=2bb65738-15dc-4488-ae8d-60a9fc3d3ea5 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:10.399496980Z" level=info msg="Starting container: fcb76a52605ae84e7415c0b6e56aa8e1f010607edc042a479cb6e4adde24ecc4" id=fd1a897b-467b-4685-8122-654b40ffcce1 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:12:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:10.425988536Z" level=info msg="Started container" PID=73160 containerID=fcb76a52605ae84e7415c0b6e56aa8e1f010607edc042a479cb6e4adde24ecc4 description=openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/thanos-query id=fd1a897b-467b-4685-8122-654b40ffcce1 name=/runtime.v1.RuntimeService/StartContainer sandboxID=6c2cdef5be97661e24b16499c5f84baa5f0fdffbceb45f19b515c50e93bb6dd5 Jan 05 09:12:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:10.493696392Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=e8442a66-ee7b-482a-b057-795a98622e69 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:10.493923804Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=e8442a66-ee7b-482a-b057-795a98622e69 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:10.519148369Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=10cdc2ae-9ee6-418f-8211-90bc5eea5791 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:10.519439739Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=10cdc2ae-9ee6-418f-8211-90bc5eea5791 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:10.524757656Z" level=info msg="Creating container: openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/oauth-proxy" id=06f92a68-ff2b-4ecd-b482-c4b20ce62ff1 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:10.524878502Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:12:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:10.726761686Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816" id=c25a47f6-ce78-4ca8-b4f6-7405b90e665d name=/runtime.v1.ImageService/PullImage Jan 05 09:12:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:10.727915505Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816" id=ef7fc247-ae23-4373-9937-ee367a5e72f5 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:10.729461159Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:955ee2b41ecd88e3a9c0183a0b65d6edbf524cefa1318837dcbc82b0466c4d55,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816],Size_:309576938,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=ef7fc247-ae23-4373-9937-ee367a5e72f5 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:10.730246957Z" level=info msg="Creating container: openshift-monitoring/alertmanager-main-1/prom-label-proxy" id=7105239c-5a44-4bc9-97ae-e9abb163f11e name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:10.730352239Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:12:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:10.806478485Z" level=info msg="Created container e67ece935b56afbeb827b9311beb6a1f9f4afc8db2691bdcd603bd752c6deaa7: openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/oauth-proxy" id=06f92a68-ff2b-4ecd-b482-c4b20ce62ff1 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:10.806919856Z" level=info msg="Starting container: e67ece935b56afbeb827b9311beb6a1f9f4afc8db2691bdcd603bd752c6deaa7" id=5baee2d8-aead-41c3-b986-1187a1c97212 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:12:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:10.840830328Z" level=info msg="Started container" PID=73214 containerID=e67ece935b56afbeb827b9311beb6a1f9f4afc8db2691bdcd603bd752c6deaa7 description=openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/oauth-proxy id=5baee2d8-aead-41c3-b986-1187a1c97212 name=/runtime.v1.RuntimeService/StartContainer sandboxID=6c2cdef5be97661e24b16499c5f84baa5f0fdffbceb45f19b515c50e93bb6dd5 Jan 05 09:12:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:10.864022330Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=126edbc1-a045-481f-b33f-e90aac61006e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:10.864235373Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=126edbc1-a045-481f-b33f-e90aac61006e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:10.868251069Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=1244adda-438a-4088-bd7f-dae792cbbf66 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:10.868476542Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=1244adda-438a-4088-bd7f-dae792cbbf66 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:10.869484958Z" level=info msg="Creating container: openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/kube-rbac-proxy" id=7f9db496-8d13-457c-993a-2c03e560e232 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:10.869609296Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:12:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:10.888147582Z" level=info msg="Created container b9bc795189fd458b4ce14c6f192418f946a1c35d2f19ab6aedd4b4185f4bf487: openshift-monitoring/alertmanager-main-1/prom-label-proxy" id=7105239c-5a44-4bc9-97ae-e9abb163f11e name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:10.888694100Z" level=info msg="Starting container: b9bc795189fd458b4ce14c6f192418f946a1c35d2f19ab6aedd4b4185f4bf487" id=3db34377-2b0f-4ea0-b115-91c8fa6dc02f name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:12:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:10.906674330Z" level=info msg="Started container" PID=73241 containerID=b9bc795189fd458b4ce14c6f192418f946a1c35d2f19ab6aedd4b4185f4bf487 description=openshift-monitoring/alertmanager-main-1/prom-label-proxy id=3db34377-2b0f-4ea0-b115-91c8fa6dc02f name=/runtime.v1.RuntimeService/StartContainer sandboxID=4cc6b8c0009aed4f5500066da7f716df203dd1969281bb8dbce4d7052e55ec50 Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:11.027293510Z" level=info msg="Created container 217b93493832653788189b59ae04c4259e18df93a8c0c89eb6d7197fdb6bdf27: openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/kube-rbac-proxy" id=7f9db496-8d13-457c-993a-2c03e560e232 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:11.027925940Z" level=info msg="Starting container: 217b93493832653788189b59ae04c4259e18df93a8c0c89eb6d7197fdb6bdf27" id=d197b0ca-20ca-46d1-84e9-d75a65afcaa5 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:11.037711486Z" level=info msg="Started container" PID=73288 containerID=217b93493832653788189b59ae04c4259e18df93a8c0c89eb6d7197fdb6bdf27 description=openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/kube-rbac-proxy id=d197b0ca-20ca-46d1-84e9-d75a65afcaa5 name=/runtime.v1.RuntimeService/StartContainer sandboxID=6c2cdef5be97661e24b16499c5f84baa5f0fdffbceb45f19b515c50e93bb6dd5 Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:11.066994531Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816" id=8df94fa1-05d0-4abc-ae41-f5767aaa7b59 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:11.068768889Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:955ee2b41ecd88e3a9c0183a0b65d6edbf524cefa1318837dcbc82b0466c4d55,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816],Size_:309576938,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=8df94fa1-05d0-4abc-ae41-f5767aaa7b59 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:11.069502373Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816" id=19a54fcd-4583-4a57-9c63-56a22ff33b4d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:11.071001741Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:955ee2b41ecd88e3a9c0183a0b65d6edbf524cefa1318837dcbc82b0466c4d55,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816],Size_:309576938,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=19a54fcd-4583-4a57-9c63-56a22ff33b4d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:11.071915485Z" level=info msg="Creating container: openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/prom-label-proxy" id=192b9572-e93b-4fbc-a55e-ffa83ad98de0 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:11.072024518Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:11.109697 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" event=&{ID:72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c Type:ContainerStarted Data:217b93493832653788189b59ae04c4259e18df93a8c0c89eb6d7197fdb6bdf27} Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:11.109733 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" event=&{ID:72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c Type:ContainerStarted Data:e67ece935b56afbeb827b9311beb6a1f9f4afc8db2691bdcd603bd752c6deaa7} Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:11.109750 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" event=&{ID:72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c Type:ContainerStarted Data:fcb76a52605ae84e7415c0b6e56aa8e1f010607edc042a479cb6e4adde24ecc4} Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:11.114586 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-1" event=&{ID:b59f4808-02df-4b2e-9694-8cf72a1b5e75 Type:ContainerStarted Data:b9bc795189fd458b4ce14c6f192418f946a1c35d2f19ab6aedd4b4185f4bf487} Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:11.259505389Z" level=info msg="Created container 81147f8c4de8241e926f4c84fcb3c6b8d438e973b16a73d5c29483986c87857b: openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/prom-label-proxy" id=192b9572-e93b-4fbc-a55e-ffa83ad98de0 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:11.260254008Z" level=info msg="Starting container: 81147f8c4de8241e926f4c84fcb3c6b8d438e973b16a73d5c29483986c87857b" id=32f94451-fc1c-4ff7-9810-2ec96b276bab name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:11.273584698Z" level=info msg="Started container" PID=73329 containerID=81147f8c4de8241e926f4c84fcb3c6b8d438e973b16a73d5c29483986c87857b description=openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/prom-label-proxy id=32f94451-fc1c-4ff7-9810-2ec96b276bab name=/runtime.v1.RuntimeService/StartContainer sandboxID=6c2cdef5be97661e24b16499c5f84baa5f0fdffbceb45f19b515c50e93bb6dd5 Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:11.287604312Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=a9aa6af4-f231-4338-a123-b0116c4ff667 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:11.287840809Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=a9aa6af4-f231-4338-a123-b0116c4ff667 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:11.288817841Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=d4818dad-1b94-402f-9d6f-aac87947eb06 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:11.288966416Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=d4818dad-1b94-402f-9d6f-aac87947eb06 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:11.289766672Z" level=info msg="Creating container: openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/kube-rbac-proxy-rules" id=eb9fda5e-4dd3-4b05-abc8-a145cf7a657d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:11.289877789Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:11.499149996Z" level=info msg="Created container aa7b352bf36356466f81f85608b01622b4f84956d931d3153417ad8abf7d95e9: openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/kube-rbac-proxy-rules" id=eb9fda5e-4dd3-4b05-abc8-a145cf7a657d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:11.499694036Z" level=info msg="Starting container: aa7b352bf36356466f81f85608b01622b4f84956d931d3153417ad8abf7d95e9" id=7e87ed07-8cc3-4bc0-8cf0-9ce73de6a575 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:11.515645984Z" level=info msg="Started container" PID=73402 containerID=aa7b352bf36356466f81f85608b01622b4f84956d931d3153417ad8abf7d95e9 description=openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/kube-rbac-proxy-rules id=7e87ed07-8cc3-4bc0-8cf0-9ce73de6a575 name=/runtime.v1.RuntimeService/StartContainer sandboxID=6c2cdef5be97661e24b16499c5f84baa5f0fdffbceb45f19b515c50e93bb6dd5 Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:11.534764940Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=57ae8ccc-674b-4411-bfd8-452bb4a437b7 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:11.535004650Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=57ae8ccc-674b-4411-bfd8-452bb4a437b7 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:11.535994853Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=eb63dab3-f6fc-4536-adab-6abf1c6dd8c0 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:11.536189938Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=eb63dab3-f6fc-4536-adab-6abf1c6dd8c0 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:11.537338497Z" level=info msg="Creating container: openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/kube-rbac-proxy-metrics" id=df260395-0c8f-410b-a9ca-f2d5c1634ee9 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:11.537471975Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:11.712568484Z" level=info msg="Created container 2bce691bc41bddf8da3b590cbdbb6b3b7604a9e34c167c7d2a14b15a944439f1: openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/kube-rbac-proxy-metrics" id=df260395-0c8f-410b-a9ca-f2d5c1634ee9 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:11.713015049Z" level=info msg="Starting container: 2bce691bc41bddf8da3b590cbdbb6b3b7604a9e34c167c7d2a14b15a944439f1" id=43ce384c-84ec-4e15-95f3-d56114f7ce10 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:12:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:11.721232517Z" level=info msg="Started container" PID=73458 containerID=2bce691bc41bddf8da3b590cbdbb6b3b7604a9e34c167c7d2a14b15a944439f1 description=openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/kube-rbac-proxy-metrics id=43ce384c-84ec-4e15-95f3-d56114f7ce10 name=/runtime.v1.RuntimeService/StartContainer sandboxID=6c2cdef5be97661e24b16499c5f84baa5f0fdffbceb45f19b515c50e93bb6dd5 Jan 05 09:12:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:12.120047 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" event=&{ID:72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c Type:ContainerStarted Data:2bce691bc41bddf8da3b590cbdbb6b3b7604a9e34c167c7d2a14b15a944439f1} Jan 05 09:12:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:12.120082 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" event=&{ID:72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c Type:ContainerStarted Data:aa7b352bf36356466f81f85608b01622b4f84956d931d3153417ad8abf7d95e9} Jan 05 09:12:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:12.120098 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" event=&{ID:72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c Type:ContainerStarted Data:81147f8c4de8241e926f4c84fcb3c6b8d438e973b16a73d5c29483986c87857b} Jan 05 09:12:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:12.120140 1425 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" Jan 05 09:12:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:15.847200 1425 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" Jan 05 09:12:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:17.901936 1425 kubelet.go:2229] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:12:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:20.853865270Z" level=warning msg="Stopping container 12ba0576c39c7251a305990e9512ca687002ea4e088fb3b78d3d7f7b77efabc5 with stop signal timed out: timeout reached after 30 seconds waiting for container process to exit" id=7b6c9e91-9144-4f2c-8e5e-a06968c8662a name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:12:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:21.149707 1425 generic.go:296] "Generic (PLEG): container finished" podID=a3cb294f-6846-4e6b-b42f-434294a1cefd containerID="12ba0576c39c7251a305990e9512ca687002ea4e088fb3b78d3d7f7b77efabc5" exitCode=137 Jan 05 09:12:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:21.149748 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-nt5pz" event=&{ID:a3cb294f-6846-4e6b-b42f-434294a1cefd Type:ContainerDied Data:12ba0576c39c7251a305990e9512ca687002ea4e088fb3b78d3d7f7b77efabc5} Jan 05 09:12:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:21.659704170Z" level=info msg="Stopped container 12ba0576c39c7251a305990e9512ca687002ea4e088fb3b78d3d7f7b77efabc5: openshift-multus/cni-sysctl-allowlist-ds-nt5pz/kube-multus-additional-cni-plugins" id=7b6c9e91-9144-4f2c-8e5e-a06968c8662a name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:12:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:21.660236697Z" level=info msg="Stopping pod sandbox: 52159b4d1e3648c7df20e17bffb14777d3a3c663e803b06565beb65f460be913" id=56ba664a-5e1f-4171-bd3a-cfcc9e2c29ed name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:12:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:21.660513056Z" level=info msg="Got pod network &{Name:cni-sysctl-allowlist-ds-nt5pz Namespace:openshift-multus ID:52159b4d1e3648c7df20e17bffb14777d3a3c663e803b06565beb65f460be913 UID:a3cb294f-6846-4e6b-b42f-434294a1cefd NetNS:/var/run/netns/6ef4c066-c3e6-40db-a7e4-24e61dbd531e Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:12:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:21.660748010Z" level=info msg="Deleting pod openshift-multus_cni-sysctl-allowlist-ds-nt5pz from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:12:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:21.687507401Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:91e7d68b28aa00e209065604b2b9ac1400671d9102581e8b49b37726acd7b6e4" id=976302a0-fd40-442e-bd3c-20729ecc0d1d name=/runtime.v1.ImageService/PullImage Jan 05 09:12:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:21.701397273Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:91e7d68b28aa00e209065604b2b9ac1400671d9102581e8b49b37726acd7b6e4" id=6bfcc792-4fd5-4e39-9501-2ac49694ebcf name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:21.704079330Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b68b0fa7ea4653b56a91a8c213d456cab2f645207af609598f9877a41585eeae,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:91e7d68b28aa00e209065604b2b9ac1400671d9102581e8b49b37726acd7b6e4],Size_:493220384,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=6bfcc792-4fd5-4e39-9501-2ac49694ebcf name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:21.706395260Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-1/prometheus" id=cae90196-e501-4fce-9920-7be995f5252b name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:21.706498847Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:12:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: 2023-01-05T09:12:21Z [verbose] Del: openshift-multus:cni-sysctl-allowlist-ds-nt5pz:a3cb294f-6846-4e6b-b42f-434294a1cefd:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:12:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:21.900596296Z" level=info msg="Stopped pod sandbox: 52159b4d1e3648c7df20e17bffb14777d3a3c663e803b06565beb65f460be913" id=56ba664a-5e1f-4171-bd3a-cfcc9e2c29ed name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:12:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:21.960253478Z" level=info msg="Created container 6becde1d0c7f5d104d26d3a1c07efaa41bf091484c482dfffd15e87e4b5bf54c: openshift-monitoring/prometheus-k8s-1/prometheus" id=cae90196-e501-4fce-9920-7be995f5252b name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:21.960766036Z" level=info msg="Starting container: 6becde1d0c7f5d104d26d3a1c07efaa41bf091484c482dfffd15e87e4b5bf54c" id=3ca52988-c1a2-4ecc-b31d-4f15a734d6a0 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:12:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:21.969417582Z" level=info msg="Started container" PID=73764 containerID=6becde1d0c7f5d104d26d3a1c07efaa41bf091484c482dfffd15e87e4b5bf54c description=openshift-monitoring/prometheus-k8s-1/prometheus id=3ca52988-c1a2-4ecc-b31d-4f15a734d6a0 name=/runtime.v1.RuntimeService/StartContainer sandboxID=bf998995ead8afb2642821b82636d7b867d8c0aa5ab164fe09e7396c2701f63b Jan 05 09:12:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:21.986503307Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=2c8fdc2a-8754-4b8d-82e5-3b3f20d80190 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:21.986752627Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=2c8fdc2a-8754-4b8d-82e5-3b3f20d80190 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:21.987800451Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=b0c9a9d7-dbcb-49c7-9a1c-e23c70936565 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:21.987990294Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b0c9a9d7-dbcb-49c7-9a1c-e23c70936565 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:21.989068181Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-1/config-reloader" id=d379ca2d-de1c-4fa0-bb11-4e32aed13612 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:21.989182339Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:22.023127 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a3cb294f-6846-4e6b-b42f-434294a1cefd-tuning-conf-dir\") pod \"a3cb294f-6846-4e6b-b42f-434294a1cefd\" (UID: \"a3cb294f-6846-4e6b-b42f-434294a1cefd\") " Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:22.023196 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a3cb294f-6846-4e6b-b42f-434294a1cefd-cni-sysctl-allowlist\") pod \"a3cb294f-6846-4e6b-b42f-434294a1cefd\" (UID: \"a3cb294f-6846-4e6b-b42f-434294a1cefd\") " Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:22.023240 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frtl7\" (UniqueName: \"kubernetes.io/projected/a3cb294f-6846-4e6b-b42f-434294a1cefd-kube-api-access-frtl7\") pod \"a3cb294f-6846-4e6b-b42f-434294a1cefd\" (UID: \"a3cb294f-6846-4e6b-b42f-434294a1cefd\") " Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:22.023281 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/a3cb294f-6846-4e6b-b42f-434294a1cefd-ready\") pod \"a3cb294f-6846-4e6b-b42f-434294a1cefd\" (UID: \"a3cb294f-6846-4e6b-b42f-434294a1cefd\") " Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:12:22.023630 1425 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/a3cb294f-6846-4e6b-b42f-434294a1cefd/volumes/kubernetes.io~empty-dir/ready: clearQuota called, but quotas disabled Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:22.023727 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3cb294f-6846-4e6b-b42f-434294a1cefd-ready" (OuterVolumeSpecName: "ready") pod "a3cb294f-6846-4e6b-b42f-434294a1cefd" (UID: "a3cb294f-6846-4e6b-b42f-434294a1cefd"). InnerVolumeSpecName "ready". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:22.023780 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a3cb294f-6846-4e6b-b42f-434294a1cefd-tuning-conf-dir" (OuterVolumeSpecName: "tuning-conf-dir") pod "a3cb294f-6846-4e6b-b42f-434294a1cefd" (UID: "a3cb294f-6846-4e6b-b42f-434294a1cefd"). InnerVolumeSpecName "tuning-conf-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:12:22.023929 1425 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/a3cb294f-6846-4e6b-b42f-434294a1cefd/volumes/kubernetes.io~configmap/cni-sysctl-allowlist: clearQuota called, but quotas disabled Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:22.024130 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3cb294f-6846-4e6b-b42f-434294a1cefd-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "a3cb294f-6846-4e6b-b42f-434294a1cefd" (UID: "a3cb294f-6846-4e6b-b42f-434294a1cefd"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:22.033094 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3cb294f-6846-4e6b-b42f-434294a1cefd-kube-api-access-frtl7" (OuterVolumeSpecName: "kube-api-access-frtl7") pod "a3cb294f-6846-4e6b-b42f-434294a1cefd" (UID: "a3cb294f-6846-4e6b-b42f-434294a1cefd"). InnerVolumeSpecName "kube-api-access-frtl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:22.123910 1425 reconciler.go:399] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a3cb294f-6846-4e6b-b42f-434294a1cefd-cni-sysctl-allowlist\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:22.123948 1425 reconciler.go:399] "Volume detached for volume \"kube-api-access-frtl7\" (UniqueName: \"kubernetes.io/projected/a3cb294f-6846-4e6b-b42f-434294a1cefd-kube-api-access-frtl7\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:22.123970 1425 reconciler.go:399] "Volume detached for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/a3cb294f-6846-4e6b-b42f-434294a1cefd-ready\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:22.123992 1425 reconciler.go:399] "Volume detached for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a3cb294f-6846-4e6b-b42f-434294a1cefd-tuning-conf-dir\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:22.153906 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-1" event=&{ID:8ac551e8-a959-4212-86a1-2b3ac482586f Type:ContainerStarted Data:6becde1d0c7f5d104d26d3a1c07efaa41bf091484c482dfffd15e87e4b5bf54c} Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:22.155505 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-nt5pz" event=&{ID:a3cb294f-6846-4e6b-b42f-434294a1cefd Type:ContainerDied Data:52159b4d1e3648c7df20e17bffb14777d3a3c663e803b06565beb65f460be913} Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:22.155566 1425 scope.go:115] "RemoveContainer" containerID="12ba0576c39c7251a305990e9512ca687002ea4e088fb3b78d3d7f7b77efabc5" Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:22.163121996Z" level=info msg="Removing container: 12ba0576c39c7251a305990e9512ca687002ea4e088fb3b78d3d7f7b77efabc5" id=23c4817f-668c-439d-8eb4-bc8dcfb336f9 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:22.197853 1425 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-multus/cni-sysctl-allowlist-ds-nt5pz] Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:22.202350 1425 kubelet.go:2129] "SyncLoop REMOVE" source="api" pods=[openshift-multus/cni-sysctl-allowlist-ds-nt5pz] Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:22.237508186Z" level=info msg="Removed container 12ba0576c39c7251a305990e9512ca687002ea4e088fb3b78d3d7f7b77efabc5: openshift-multus/cni-sysctl-allowlist-ds-nt5pz/kube-multus-additional-cni-plugins" id=23c4817f-668c-439d-8eb4-bc8dcfb336f9 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:22.281129237Z" level=info msg="Created container 254f6adadc261bc8b8bef366241b799f8c167b2d7c165cb79c16c15ee634f029: openshift-monitoring/prometheus-k8s-1/config-reloader" id=d379ca2d-de1c-4fa0-bb11-4e32aed13612 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:22.293174673Z" level=info msg="Starting container: 254f6adadc261bc8b8bef366241b799f8c167b2d7c165cb79c16c15ee634f029" id=42505555-12e9-4383-98a1-05d2fe2f2099 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:22.315895320Z" level=info msg="Started container" PID=73805 containerID=254f6adadc261bc8b8bef366241b799f8c167b2d7c165cb79c16c15ee634f029 description=openshift-monitoring/prometheus-k8s-1/config-reloader id=42505555-12e9-4383-98a1-05d2fe2f2099 name=/runtime.v1.RuntimeService/StartContainer sandboxID=bf998995ead8afb2642821b82636d7b867d8c0aa5ab164fe09e7396c2701f63b Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:22.362006008Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc" id=25a8c849-46a3-4921-8ee7-517c84d840fb name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:22.362236457Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:08ff7a43d9185529520a29e114f0c28b070000b464a44b347240015f3b466488,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc],Size_:368278089,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=25a8c849-46a3-4921-8ee7-517c84d840fb name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:22.363182691Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc" id=fce10abe-f0af-4700-8641-3e5b80bdd1b0 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:22.363437388Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:08ff7a43d9185529520a29e114f0c28b070000b464a44b347240015f3b466488,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc],Size_:368278089,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=fce10abe-f0af-4700-8641-3e5b80bdd1b0 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:22.364522597Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-1/thanos-sidecar" id=344470fa-ed08-46be-8ce3-c9f14d7374cd name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:22.364654128Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:22.669620 1425 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=a3cb294f-6846-4e6b-b42f-434294a1cefd path="/var/lib/kubelet/pods/a3cb294f-6846-4e6b-b42f-434294a1cefd/volumes" Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:22.733657719Z" level=info msg="Created container 976ee1fd7cc0d02f224dbcf7a41285bac05287d0f5cf4e79c3f6c098c2a7dcd4: openshift-monitoring/prometheus-k8s-1/thanos-sidecar" id=344470fa-ed08-46be-8ce3-c9f14d7374cd name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:22.734303952Z" level=info msg="Starting container: 976ee1fd7cc0d02f224dbcf7a41285bac05287d0f5cf4e79c3f6c098c2a7dcd4" id=52c114ba-3070-4482-8a25-8b601c47795f name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:22.773947253Z" level=info msg="Started container" PID=73854 containerID=976ee1fd7cc0d02f224dbcf7a41285bac05287d0f5cf4e79c3f6c098c2a7dcd4 description=openshift-monitoring/prometheus-k8s-1/thanos-sidecar id=52c114ba-3070-4482-8a25-8b601c47795f name=/runtime.v1.RuntimeService/StartContainer sandboxID=bf998995ead8afb2642821b82636d7b867d8c0aa5ab164fe09e7396c2701f63b Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:22.798699635Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=5b67dec8-24bf-4109-94e1-3a5b4a906f8b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:22.798921575Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=5b67dec8-24bf-4109-94e1-3a5b4a906f8b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:22.800015962Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=163ef1bd-7410-49fe-9078-3da1682cd210 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:22.800206887Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=163ef1bd-7410-49fe-9078-3da1682cd210 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:22.801778721Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-1/prometheus-proxy" id=0b404e3b-236f-41b3-8c18-0e484c1d35bd name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:22.801890003Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:12:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:23.175788 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-1" event=&{ID:8ac551e8-a959-4212-86a1-2b3ac482586f Type:ContainerStarted Data:976ee1fd7cc0d02f224dbcf7a41285bac05287d0f5cf4e79c3f6c098c2a7dcd4} Jan 05 09:12:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:23.175823 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-1" event=&{ID:8ac551e8-a959-4212-86a1-2b3ac482586f Type:ContainerStarted Data:254f6adadc261bc8b8bef366241b799f8c167b2d7c165cb79c16c15ee634f029} Jan 05 09:12:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:23.182673665Z" level=info msg="Created container e6208ed7b9782cddc95b22490d2dc4f820f1944cb39050a5d478fabf458cd570: openshift-monitoring/prometheus-k8s-1/prometheus-proxy" id=0b404e3b-236f-41b3-8c18-0e484c1d35bd name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:23.183618486Z" level=info msg="Starting container: e6208ed7b9782cddc95b22490d2dc4f820f1944cb39050a5d478fabf458cd570" id=80d4dc76-d65a-4882-b5f3-2f528ad49741 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:12:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:23.213781874Z" level=info msg="Started container" PID=73896 containerID=e6208ed7b9782cddc95b22490d2dc4f820f1944cb39050a5d478fabf458cd570 description=openshift-monitoring/prometheus-k8s-1/prometheus-proxy id=80d4dc76-d65a-4882-b5f3-2f528ad49741 name=/runtime.v1.RuntimeService/StartContainer sandboxID=bf998995ead8afb2642821b82636d7b867d8c0aa5ab164fe09e7396c2701f63b Jan 05 09:12:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:23.232914959Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=6f5f543c-0bc9-4bbe-8398-cfbe1a004925 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:23.233164743Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=6f5f543c-0bc9-4bbe-8398-cfbe1a004925 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:23.234036666Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=239882ce-681d-4ca3-8b78-2075cc601cae name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:23.234231907Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=239882ce-681d-4ca3-8b78-2075cc601cae name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:23.235294006Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-1/kube-rbac-proxy" id=614c3c41-23d0-4507-ade4-5356371c8241 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:23.235436252Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:12:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:23.372400923Z" level=info msg="Created container 67e4592536523011a4901ffad309428f68583ceb75cd44ec7d261c76a0778640: openshift-monitoring/prometheus-k8s-1/kube-rbac-proxy" id=614c3c41-23d0-4507-ade4-5356371c8241 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:23.373019980Z" level=info msg="Starting container: 67e4592536523011a4901ffad309428f68583ceb75cd44ec7d261c76a0778640" id=11c3afbd-03f3-410c-ab06-5cb14fdfc892 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:12:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:23.391266720Z" level=info msg="Started container" PID=73933 containerID=67e4592536523011a4901ffad309428f68583ceb75cd44ec7d261c76a0778640 description=openshift-monitoring/prometheus-k8s-1/kube-rbac-proxy id=11c3afbd-03f3-410c-ab06-5cb14fdfc892 name=/runtime.v1.RuntimeService/StartContainer sandboxID=bf998995ead8afb2642821b82636d7b867d8c0aa5ab164fe09e7396c2701f63b Jan 05 09:12:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:23.408618429Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=d0be8d37-adc8-4342-96c1-b21f2f6ef517 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:23.408857685Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=d0be8d37-adc8-4342-96c1-b21f2f6ef517 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:23.409950060Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=a0f3c473-66ae-46c5-be45-91f0d9b8c49a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:23.410156558Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=a0f3c473-66ae-46c5-be45-91f0d9b8c49a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:23.411353767Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-1/kube-rbac-proxy-thanos" id=fa3fc68d-04ec-46ca-bea1-25d2e56cf1b6 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:23.411494458Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:12:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:23.512600539Z" level=info msg="Created container 1a898502c172251a58f921f830b7d6d6f8f096259c3dca6353d228972a83e528: openshift-monitoring/prometheus-k8s-1/kube-rbac-proxy-thanos" id=fa3fc68d-04ec-46ca-bea1-25d2e56cf1b6 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:12:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:23.513041042Z" level=info msg="Starting container: 1a898502c172251a58f921f830b7d6d6f8f096259c3dca6353d228972a83e528" id=141f7981-5653-43f9-956b-ffc7e890d538 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:12:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:23.520903587Z" level=info msg="Started container" PID=73971 containerID=1a898502c172251a58f921f830b7d6d6f8f096259c3dca6353d228972a83e528 description=openshift-monitoring/prometheus-k8s-1/kube-rbac-proxy-thanos id=141f7981-5653-43f9-956b-ffc7e890d538 name=/runtime.v1.RuntimeService/StartContainer sandboxID=bf998995ead8afb2642821b82636d7b867d8c0aa5ab164fe09e7396c2701f63b Jan 05 09:12:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:24.182650 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-1" event=&{ID:8ac551e8-a959-4212-86a1-2b3ac482586f Type:ContainerStarted Data:1a898502c172251a58f921f830b7d6d6f8f096259c3dca6353d228972a83e528} Jan 05 09:12:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:24.182687 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-1" event=&{ID:8ac551e8-a959-4212-86a1-2b3ac482586f Type:ContainerStarted Data:67e4592536523011a4901ffad309428f68583ceb75cd44ec7d261c76a0778640} Jan 05 09:12:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:24.182702 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-1" event=&{ID:8ac551e8-a959-4212-86a1-2b3ac482586f Type:ContainerStarted Data:e6208ed7b9782cddc95b22490d2dc4f820f1944cb39050a5d478fabf458cd570} Jan 05 09:12:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:27.942485 1425 kubelet.go:2229] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:12:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:28.919613 1425 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:12:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:30.177441 1425 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" Jan 05 09:12:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:36.480302284Z" level=warning msg="Found defunct process with PID 73592 (haproxy)" Jan 05 09:12:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:36.505797599Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=f78b2506-0faf-4187-b59b-4b37d177f737 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:36.506041103Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=f78b2506-0faf-4187-b59b-4b37d177f737 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:12:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:36.658242731Z" level=info msg="Stopping pod sandbox: 52159b4d1e3648c7df20e17bffb14777d3a3c663e803b06565beb65f460be913" id=5f1256aa-9f3e-404b-b413-921b0904ce2c name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:12:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:36.658284667Z" level=info msg="Stopped pod sandbox (already stopped): 52159b4d1e3648c7df20e17bffb14777d3a3c663e803b06565beb65f460be913" id=5f1256aa-9f3e-404b-b413-921b0904ce2c name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:12:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:36.658640584Z" level=info msg="Removing pod sandbox: 52159b4d1e3648c7df20e17bffb14777d3a3c663e803b06565beb65f460be913" id=788b16c8-c29d-4358-a5e1-41af9506ec3a name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:12:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:12:36.667223107Z" level=info msg="Removed pod sandbox: 52159b4d1e3648c7df20e17bffb14777d3a3c663e803b06565beb65f460be913" id=788b16c8-c29d-4358-a5e1-41af9506ec3a name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:12:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:38.918734 1425 kubelet.go:2229] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:12:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:38.962277 1425 kubelet.go:2229] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:12:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:12:39.277705 1425 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.296628 1425 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-network-diagnostics/network-check-source-6775957967-zsr9c] Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.296828 1425 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-network-diagnostics/network-check-source-6775957967-zsr9c" podUID=89a48689-810c-464e-9ac6-d78977e60793 containerName="check-endpoints" containerID="cri-o://3b8c93b35afc863ee93d43a145f66340ca0ab0a562aca1259bff690246d320cf" gracePeriod=30 Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:10.297345242Z" level=info msg="Stopping container: 3b8c93b35afc863ee93d43a145f66340ca0ab0a562aca1259bff690246d320cf (timeout: 30s)" id=083f8f8a-768c-44da-bd35-3f3076cd42d2 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.308047 1425 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-ingress/router-default-5cf6bbdff9-fkmcn] Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.308253 1425 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" podUID=66d22b70-08bb-4199-b258-ac4590e9b487 containerName="router" containerID="cri-o://8c886b3e9fbca71166a4d01928168d286331774947ddef0326c331f3624f5f6a" gracePeriod=3600 Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:10.311666906Z" level=info msg="Stopping container: 8c886b3e9fbca71166a4d01928168d286331774947ddef0326c331f3624f5f6a (timeout: 3600s)" id=9c26b5a4-93b8-4eb7-ba4a-6d69c5d1f52d name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.329051 1425 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-monitoring/alertmanager-main-1] Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.329093 1425 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-54rjq] Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.329113 1425 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-monitoring/prometheus-k8s-1] Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.329355 1425 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/prometheus-k8s-1" podUID=8ac551e8-a959-4212-86a1-2b3ac482586f containerName="prometheus" containerID="cri-o://6becde1d0c7f5d104d26d3a1c07efaa41bf091484c482dfffd15e87e4b5bf54c" gracePeriod=600 Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.329651 1425 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-1" podUID=b59f4808-02df-4b2e-9694-8cf72a1b5e75 containerName="alertmanager" containerID="cri-o://902fedd514d362305b82903fec325b5adb139bd729665753e388c28b1e4d58bf" gracePeriod=120 Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.329794 1425 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-54rjq" podUID=f8e5be3f-9ab7-484c-b61b-6231a8774e24 containerName="prometheus-operator-admission-webhook" containerID="cri-o://b6613cda35aa605c0c2f9fa03950bbd6def356ba0a1a95a665f8c80d6b17a14f" gracePeriod=30 Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.329879 1425 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/prometheus-k8s-1" podUID=8ac551e8-a959-4212-86a1-2b3ac482586f containerName="kube-rbac-proxy-thanos" containerID="cri-o://1a898502c172251a58f921f830b7d6d6f8f096259c3dca6353d228972a83e528" gracePeriod=600 Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.329932 1425 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/prometheus-k8s-1" podUID=8ac551e8-a959-4212-86a1-2b3ac482586f containerName="kube-rbac-proxy" containerID="cri-o://67e4592536523011a4901ffad309428f68583ceb75cd44ec7d261c76a0778640" gracePeriod=600 Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.329983 1425 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/prometheus-k8s-1" podUID=8ac551e8-a959-4212-86a1-2b3ac482586f containerName="prometheus-proxy" containerID="cri-o://e6208ed7b9782cddc95b22490d2dc4f820f1944cb39050a5d478fabf458cd570" gracePeriod=600 Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.330033 1425 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/prometheus-k8s-1" podUID=8ac551e8-a959-4212-86a1-2b3ac482586f containerName="thanos-sidecar" containerID="cri-o://976ee1fd7cc0d02f224dbcf7a41285bac05287d0f5cf4e79c3f6c098c2a7dcd4" gracePeriod=600 Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.330087 1425 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/prometheus-k8s-1" podUID=8ac551e8-a959-4212-86a1-2b3ac482586f containerName="config-reloader" containerID="cri-o://254f6adadc261bc8b8bef366241b799f8c167b2d7c165cb79c16c15ee634f029" gracePeriod=600 Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.330200 1425 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-1" podUID=b59f4808-02df-4b2e-9694-8cf72a1b5e75 containerName="prom-label-proxy" containerID="cri-o://b9bc795189fd458b4ce14c6f192418f946a1c35d2f19ab6aedd4b4185f4bf487" gracePeriod=120 Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.330260 1425 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-1" podUID=b59f4808-02df-4b2e-9694-8cf72a1b5e75 containerName="kube-rbac-proxy-metric" containerID="cri-o://064809cddb0902a863e4fcd8d1da29b6a436aa3118d40b3b5ab6fbbda4fbf3d9" gracePeriod=120 Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.330313 1425 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-1" podUID=b59f4808-02df-4b2e-9694-8cf72a1b5e75 containerName="kube-rbac-proxy" containerID="cri-o://6eba543bd270671b29706b350f4dfdf034230e4d33c811838d47249838c4ff2b" gracePeriod=120 Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.330362 1425 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-1" podUID=b59f4808-02df-4b2e-9694-8cf72a1b5e75 containerName="alertmanager-proxy" containerID="cri-o://af5d20453c1aa7ea8d99d15e82767c98893d319f38bb9cd97aa815202e28d32c" gracePeriod=120 Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.330446 1425 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-1" podUID=b59f4808-02df-4b2e-9694-8cf72a1b5e75 containerName="config-reloader" containerID="cri-o://e9efd1bb280b0736df20aa595c1c6f65c0d953e8a7a6a1dde105be0f728d6de6" gracePeriod=120 Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:10.340823964Z" level=info msg="Stopping container: e9efd1bb280b0736df20aa595c1c6f65c0d953e8a7a6a1dde105be0f728d6de6 (timeout: 120s)" id=357de195-5f82-4bab-a383-ff3e7bcb43fb name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:10.344487227Z" level=info msg="Stopping container: e6208ed7b9782cddc95b22490d2dc4f820f1944cb39050a5d478fabf458cd570 (timeout: 600s)" id=3656f1dd-862d-4b10-9fa0-2d21886d01a8 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:10.350110744Z" level=info msg="Stopping container: 976ee1fd7cc0d02f224dbcf7a41285bac05287d0f5cf4e79c3f6c098c2a7dcd4 (timeout: 600s)" id=1dea3b99-2d00-4195-ad3e-ec343b2a0a6d name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.353191 1425 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-kube-storage-version-migrator/migrator-6cd87bdff8-p9dkf] Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.353235 1425 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-image-registry/image-registry-64655f6865-dx2q6] Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.353623 1425 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-kube-storage-version-migrator/migrator-6cd87bdff8-p9dkf" podUID=26f1eb05-de00-4fc2-864c-6b97e7f82e96 containerName="migrator" containerID="cri-o://ff17947445f945fe096ba0ee1a903135c1548bf46b79e37353e7eddd5042f6f8" gracePeriod=30 Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:10.361650567Z" level=info msg="Stopping container: 6becde1d0c7f5d104d26d3a1c07efaa41bf091484c482dfffd15e87e4b5bf54c (timeout: 600s)" id=396f1a7e-445e-41ae-ad29-4f928e58d02b name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:10.361804027Z" level=info msg="Stopping container: 902fedd514d362305b82903fec325b5adb139bd729665753e388c28b1e4d58bf (timeout: 120s)" id=ea616b09-e5b6-49a7-b14a-5fc815e347d1 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:10.361921291Z" level=info msg="Stopping container: b6613cda35aa605c0c2f9fa03950bbd6def356ba0a1a95a665f8c80d6b17a14f (timeout: 30s)" id=900acb82-ab4d-4a8b-96d0-e486d64b08d8 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:10.362044332Z" level=info msg="Stopping container: 1a898502c172251a58f921f830b7d6d6f8f096259c3dca6353d228972a83e528 (timeout: 600s)" id=38ba3ff4-e211-47bb-9c5c-b281cf8c00a6 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:10.362143773Z" level=info msg="Stopping container: 67e4592536523011a4901ffad309428f68583ceb75cd44ec7d261c76a0778640 (timeout: 600s)" id=ce076068-c46a-4e36-95a8-678dc15923c4 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:10.362309037Z" level=info msg="Stopping container: ff17947445f945fe096ba0ee1a903135c1548bf46b79e37353e7eddd5042f6f8 (timeout: 30s)" id=d87ca56a-6643-43c1-8581-c3b1b48327d1 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:10.375926292Z" level=info msg="Stopping container: 254f6adadc261bc8b8bef366241b799f8c167b2d7c165cb79c16c15ee634f029 (timeout: 600s)" id=3010c1aa-9971-4d15-b527-64fe20b46f0d name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:10.376017644Z" level=info msg="Stopping container: b9bc795189fd458b4ce14c6f192418f946a1c35d2f19ab6aedd4b4185f4bf487 (timeout: 120s)" id=5c61f5b8-1eab-4388-8697-99e64a64a25b name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.377619 1425 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-monitoring/prometheus-adapter-5894cc86c7-47422] Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.377655 1425 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx] Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.377864 1425 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" podUID=72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c containerName="thanos-query" containerID="cri-o://fcb76a52605ae84e7415c0b6e56aa8e1f010607edc042a479cb6e4adde24ecc4" gracePeriod=120 Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.378050 1425 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" podUID=a82fd83b-5344-4e7a-91dd-5894f3846e87 containerName="prometheus-adapter" containerID="cri-o://a30b14e5cf7f780de074100612e1db98e849c8b763e7b551ab0f6558754f69f6" gracePeriod=30 Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.378119 1425 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" podUID=72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c containerName="kube-rbac-proxy-metrics" containerID="cri-o://2bce691bc41bddf8da3b590cbdbb6b3b7604a9e34c167c7d2a14b15a944439f1" gracePeriod=120 Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.378171 1425 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" podUID=72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c containerName="kube-rbac-proxy-rules" containerID="cri-o://aa7b352bf36356466f81f85608b01622b4f84956d931d3153417ad8abf7d95e9" gracePeriod=120 Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.378223 1425 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" podUID=72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c containerName="prom-label-proxy" containerID="cri-o://81147f8c4de8241e926f4c84fcb3c6b8d438e973b16a73d5c29483986c87857b" gracePeriod=120 Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.378273 1425 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" podUID=72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c containerName="kube-rbac-proxy" containerID="cri-o://217b93493832653788189b59ae04c4259e18df93a8c0c89eb6d7197fdb6bdf27" gracePeriod=120 Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.378322 1425 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" podUID=72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c containerName="oauth-proxy" containerID="cri-o://e67ece935b56afbeb827b9311beb6a1f9f4afc8db2691bdcd603bd752c6deaa7" gracePeriod=120 Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:10.379175273Z" level=info msg="Stopping container: 064809cddb0902a863e4fcd8d1da29b6a436aa3118d40b3b5ab6fbbda4fbf3d9 (timeout: 120s)" id=377ca941-121b-46e2-8408-e511dc5fc965 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:10.380946970Z" level=info msg="Stopping container: 6eba543bd270671b29706b350f4dfdf034230e4d33c811838d47249838c4ff2b (timeout: 120s)" id=eb1a3527-d01d-4b23-b2c8-c77137b0745e name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:10.381066685Z" level=info msg="Stopping container: af5d20453c1aa7ea8d99d15e82767c98893d319f38bb9cd97aa815202e28d32c (timeout: 120s)" id=08b41ca9-4d5a-48db-afa9-3b246ee95e5c name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:10.381442972Z" level=info msg="Stopping container: e67ece935b56afbeb827b9311beb6a1f9f4afc8db2691bdcd603bd752c6deaa7 (timeout: 120s)" id=686ecdd5-e183-48ff-b0af-359a9cecb471 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:10.381661361Z" level=info msg="Stopping container: fcb76a52605ae84e7415c0b6e56aa8e1f010607edc042a479cb6e4adde24ecc4 (timeout: 120s)" id=1f769a43-abf8-4247-8df8-48ae094a9e4d name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:10.381786201Z" level=info msg="Stopping container: a30b14e5cf7f780de074100612e1db98e849c8b763e7b551ab0f6558754f69f6 (timeout: 30s)" id=7237ed66-7e91-4768-9e9c-1d49d92e0b75 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:10.381905826Z" level=info msg="Stopping container: 2bce691bc41bddf8da3b590cbdbb6b3b7604a9e34c167c7d2a14b15a944439f1 (timeout: 120s)" id=c087e6b1-94c4-4693-abec-7717fc3a4ba5 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:10.382015105Z" level=info msg="Stopping container: aa7b352bf36356466f81f85608b01622b4f84956d931d3153417ad8abf7d95e9 (timeout: 120s)" id=cc7e7c19-17c5-42b1-9497-34a880b5d5bb name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:10.382129730Z" level=info msg="Stopping container: 81147f8c4de8241e926f4c84fcb3c6b8d438e973b16a73d5c29483986c87857b (timeout: 120s)" id=e53fcbfa-9e0e-443b-938b-4fdb6bf644b8 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:10.382238667Z" level=info msg="Stopping container: 217b93493832653788189b59ae04c4259e18df93a8c0c89eb6d7197fdb6bdf27 (timeout: 120s)" id=d362f2b7-90dc-4cf5-bbdf-2c528809bbeb name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.845232 1425 patch_prober.go:29] interesting pod/thanos-querier-6f4f5fb7cf-wtcwx container/oauth-proxy namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.129.2.11:9091/-/ready\": dial tcp 10.129.2.11:9091: connect: connection refused" start-of-body= Jan 05 09:14:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:10.845296 1425 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" podUID=72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c containerName="oauth-proxy" probeResult=failure output="Get \"https://10.129.2.11:9091/-/ready\": dial tcp 10.129.2.11:9091: connect: connection refused" Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.045249755Z" level=info msg="Stopped container ff17947445f945fe096ba0ee1a903135c1548bf46b79e37353e7eddd5042f6f8: openshift-kube-storage-version-migrator/migrator-6cd87bdff8-p9dkf/migrator" id=d87ca56a-6643-43c1-8581-c3b1b48327d1 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.050204235Z" level=info msg="Stopping pod sandbox: 148c6bc4da65fd9b07c5fca9d24c7096ee2b924f8d407958ac234e9758e340f4" id=8f04cfe9-6207-4b72-8dd9-17a88a7fd6a1 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.050490874Z" level=info msg="Got pod network &{Name:migrator-6cd87bdff8-p9dkf Namespace:openshift-kube-storage-version-migrator ID:148c6bc4da65fd9b07c5fca9d24c7096ee2b924f8d407958ac234e9758e340f4 UID:26f1eb05-de00-4fc2-864c-6b97e7f82e96 NetNS:/var/run/netns/7b49749d-6d7c-4bf9-bed4-8bb736f61214 Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.050651432Z" level=info msg="Deleting pod openshift-kube-storage-version-migrator_migrator-6cd87bdff8-p9dkf from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.146515993Z" level=info msg="Stopped container 3b8c93b35afc863ee93d43a145f66340ca0ab0a562aca1259bff690246d320cf: openshift-network-diagnostics/network-check-source-6775957967-zsr9c/check-endpoints" id=083f8f8a-768c-44da-bd35-3f3076cd42d2 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.149866719Z" level=info msg="Stopping pod sandbox: e7532666db14128bc848af4cbb52adc603db1ec585b807063b42799df403a10b" id=f32a029b-8b29-4c82-aaea-600888a0ebbb name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.150117989Z" level=info msg="Got pod network &{Name:network-check-source-6775957967-zsr9c Namespace:openshift-network-diagnostics ID:e7532666db14128bc848af4cbb52adc603db1ec585b807063b42799df403a10b UID:89a48689-810c-464e-9ac6-d78977e60793 NetNS:/var/run/netns/61898809-3b8f-4a3e-bbd7-8989e7ede22e Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.150258072Z" level=info msg="Deleting pod openshift-network-diagnostics_network-check-source-6775957967-zsr9c from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.192163255Z" level=info msg="Stopped container b6613cda35aa605c0c2f9fa03950bbd6def356ba0a1a95a665f8c80d6b17a14f: openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-54rjq/prometheus-operator-admission-webhook" id=900acb82-ab4d-4a8b-96d0-e486d64b08d8 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.196616737Z" level=info msg="Stopping pod sandbox: 8e21977f4f2aa77ecd7154452d4a15adc91d3b323c169dda297ad9c1f8901298" id=01f05c26-4ddf-4ed0-a1ec-c03c22a9e8d9 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.196869315Z" level=info msg="Got pod network &{Name:prometheus-operator-admission-webhook-598468bb49-54rjq Namespace:openshift-monitoring ID:8e21977f4f2aa77ecd7154452d4a15adc91d3b323c169dda297ad9c1f8901298 UID:f8e5be3f-9ab7-484c-b61b-6231a8774e24 NetNS:/var/run/netns/7ace1278-51dd-403a-b0d3-915a3ab684a5 Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.197015433Z" level=info msg="Deleting pod openshift-monitoring_prometheus-operator-admission-webhook-598468bb49-54rjq from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.201371600Z" level=info msg="Stopped container e67ece935b56afbeb827b9311beb6a1f9f4afc8db2691bdcd603bd752c6deaa7: openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/oauth-proxy" id=686ecdd5-e183-48ff-b0af-359a9cecb471 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.218136544Z" level=info msg="Stopped container af5d20453c1aa7ea8d99d15e82767c98893d319f38bb9cd97aa815202e28d32c: openshift-monitoring/alertmanager-main-1/alertmanager-proxy" id=08b41ca9-4d5a-48db-afa9-3b246ee95e5c name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.226376218Z" level=info msg="Stopped container 6becde1d0c7f5d104d26d3a1c07efaa41bf091484c482dfffd15e87e4b5bf54c: openshift-monitoring/prometheus-k8s-1/prometheus" id=396f1a7e-445e-41ae-ad29-4f928e58d02b name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.248025672Z" level=info msg="Stopped container 217b93493832653788189b59ae04c4259e18df93a8c0c89eb6d7197fdb6bdf27: openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/kube-rbac-proxy" id=d362f2b7-90dc-4cf5-bbdf-2c528809bbeb name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.281700185Z" level=info msg="Stopped container 81147f8c4de8241e926f4c84fcb3c6b8d438e973b16a73d5c29483986c87857b: openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/prom-label-proxy" id=e53fcbfa-9e0e-443b-938b-4fdb6bf644b8 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.342431430Z" level=info msg="Stopped container 976ee1fd7cc0d02f224dbcf7a41285bac05287d0f5cf4e79c3f6c098c2a7dcd4: openshift-monitoring/prometheus-k8s-1/thanos-sidecar" id=1dea3b99-2d00-4195-ad3e-ec343b2a0a6d name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.386286859Z" level=info msg="Stopped container e6208ed7b9782cddc95b22490d2dc4f820f1944cb39050a5d478fabf458cd570: openshift-monitoring/prometheus-k8s-1/prometheus-proxy" id=3656f1dd-862d-4b10-9fa0-2d21886d01a8 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.426125481Z" level=info msg="Stopped container fcb76a52605ae84e7415c0b6e56aa8e1f010607edc042a479cb6e4adde24ecc4: openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/thanos-query" id=1f769a43-abf8-4247-8df8-48ae094a9e4d name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.438191684Z" level=info msg="Stopped container e9efd1bb280b0736df20aa595c1c6f65c0d953e8a7a6a1dde105be0f728d6de6: openshift-monitoring/alertmanager-main-1/config-reloader" id=357de195-5f82-4bab-a383-ff3e7bcb43fb name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.452004168Z" level=info msg="Stopped container b9bc795189fd458b4ce14c6f192418f946a1c35d2f19ab6aedd4b4185f4bf487: openshift-monitoring/alertmanager-main-1/prom-label-proxy" id=5c61f5b8-1eab-4388-8697-99e64a64a25b name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: 2023-01-05T09:14:11Z [verbose] Del: openshift-kube-storage-version-migrator:migrator-6cd87bdff8-p9dkf:26f1eb05-de00-4fc2-864c-6b97e7f82e96:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.469782269Z" level=info msg="Stopped container 6eba543bd270671b29706b350f4dfdf034230e4d33c811838d47249838c4ff2b: openshift-monitoring/alertmanager-main-1/kube-rbac-proxy" id=eb1a3527-d01d-4b23-b2c8-c77137b0745e name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.492917431Z" level=info msg="Stopped container aa7b352bf36356466f81f85608b01622b4f84956d931d3153417ad8abf7d95e9: openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/kube-rbac-proxy-rules" id=cc7e7c19-17c5-42b1-9497-34a880b5d5bb name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.506133465Z" level=info msg="Stopped container 254f6adadc261bc8b8bef366241b799f8c167b2d7c165cb79c16c15ee634f029: openshift-monitoring/prometheus-k8s-1/config-reloader" id=3010c1aa-9971-4d15-b527-64fe20b46f0d name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.515925494Z" level=info msg="Stopped container 902fedd514d362305b82903fec325b5adb139bd729665753e388c28b1e4d58bf: openshift-monitoring/alertmanager-main-1/alertmanager" id=ea616b09-e5b6-49a7-b14a-5fc815e347d1 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: 2023-01-05T09:14:11Z [verbose] Del: openshift-network-diagnostics:network-check-source-6775957967-zsr9c:89a48689-810c-464e-9ac6-d78977e60793:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.554738517Z" level=info msg="Stopped pod sandbox: 148c6bc4da65fd9b07c5fca9d24c7096ee2b924f8d407958ac234e9758e340f4" id=8f04cfe9-6207-4b72-8dd9-17a88a7fd6a1 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.573699 1425 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator_migrator-6cd87bdff8-p9dkf_26f1eb05-de00-4fc2-864c-6b97e7f82e96/migrator/0.log" Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.639030 1425 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator_migrator-6cd87bdff8-p9dkf_26f1eb05-de00-4fc2-864c-6b97e7f82e96/migrator/0.log" Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.639086 1425 generic.go:296] "Generic (PLEG): container finished" podID=26f1eb05-de00-4fc2-864c-6b97e7f82e96 containerID="ff17947445f945fe096ba0ee1a903135c1548bf46b79e37353e7eddd5042f6f8" exitCode=2 Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.639152 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-6cd87bdff8-p9dkf" event=&{ID:26f1eb05-de00-4fc2-864c-6b97e7f82e96 Type:ContainerDied Data:ff17947445f945fe096ba0ee1a903135c1548bf46b79e37353e7eddd5042f6f8} Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.639186 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-6cd87bdff8-p9dkf" event=&{ID:26f1eb05-de00-4fc2-864c-6b97e7f82e96 Type:ContainerDied Data:148c6bc4da65fd9b07c5fca9d24c7096ee2b924f8d407958ac234e9758e340f4} Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.639212 1425 scope.go:115] "RemoveContainer" containerID="ff17947445f945fe096ba0ee1a903135c1548bf46b79e37353e7eddd5042f6f8" Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.641434287Z" level=info msg="Removing container: ff17947445f945fe096ba0ee1a903135c1548bf46b79e37353e7eddd5042f6f8" id=8fcae286-e8c5-4e03-aa63-e13ff3deaf5f name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.651271706Z" level=info msg="Stopped pod sandbox: e7532666db14128bc848af4cbb52adc603db1ec585b807063b42799df403a10b" id=f32a029b-8b29-4c82-aaea-600888a0ebbb name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.652790 1425 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-6f4f5fb7cf-wtcwx_72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c/oauth-proxy/0.log" Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.654017 1425 generic.go:296] "Generic (PLEG): container finished" podID=72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c containerID="aa7b352bf36356466f81f85608b01622b4f84956d931d3153417ad8abf7d95e9" exitCode=0 Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.654038 1425 generic.go:296] "Generic (PLEG): container finished" podID=72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c containerID="81147f8c4de8241e926f4c84fcb3c6b8d438e973b16a73d5c29483986c87857b" exitCode=0 Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.654055 1425 generic.go:296] "Generic (PLEG): container finished" podID=72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c containerID="217b93493832653788189b59ae04c4259e18df93a8c0c89eb6d7197fdb6bdf27" exitCode=0 Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.654070 1425 generic.go:296] "Generic (PLEG): container finished" podID=72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c containerID="e67ece935b56afbeb827b9311beb6a1f9f4afc8db2691bdcd603bd752c6deaa7" exitCode=2 Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.654085 1425 generic.go:296] "Generic (PLEG): container finished" podID=72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c containerID="fcb76a52605ae84e7415c0b6e56aa8e1f010607edc042a479cb6e4adde24ecc4" exitCode=0 Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.654133 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" event=&{ID:72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c Type:ContainerDied Data:aa7b352bf36356466f81f85608b01622b4f84956d931d3153417ad8abf7d95e9} Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.654157 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" event=&{ID:72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c Type:ContainerDied Data:81147f8c4de8241e926f4c84fcb3c6b8d438e973b16a73d5c29483986c87857b} Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.654174 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" event=&{ID:72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c Type:ContainerDied Data:217b93493832653788189b59ae04c4259e18df93a8c0c89eb6d7197fdb6bdf27} Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.654191 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" event=&{ID:72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c Type:ContainerDied Data:e67ece935b56afbeb827b9311beb6a1f9f4afc8db2691bdcd603bd752c6deaa7} Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.654206 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" event=&{ID:72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c Type:ContainerDied Data:fcb76a52605ae84e7415c0b6e56aa8e1f010607edc042a479cb6e4adde24ecc4} Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.655651 1425 generic.go:296] "Generic (PLEG): container finished" podID=f8e5be3f-9ab7-484c-b61b-6231a8774e24 containerID="b6613cda35aa605c0c2f9fa03950bbd6def356ba0a1a95a665f8c80d6b17a14f" exitCode=0 Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.655702 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-54rjq" event=&{ID:f8e5be3f-9ab7-484c-b61b-6231a8774e24 Type:ContainerDied Data:b6613cda35aa605c0c2f9fa03950bbd6def356ba0a1a95a665f8c80d6b17a14f} Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.657580 1425 generic.go:296] "Generic (PLEG): container finished" podID=89a48689-810c-464e-9ac6-d78977e60793 containerID="3b8c93b35afc863ee93d43a145f66340ca0ab0a562aca1259bff690246d320cf" exitCode=0 Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.657626 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-6775957967-zsr9c" event=&{ID:89a48689-810c-464e-9ac6-d78977e60793 Type:ContainerDied Data:3b8c93b35afc863ee93d43a145f66340ca0ab0a562aca1259bff690246d320cf} Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.660774 1425 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-1_b59f4808-02df-4b2e-9694-8cf72a1b5e75/alertmanager-proxy/0.log" Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.661116 1425 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-1_b59f4808-02df-4b2e-9694-8cf72a1b5e75/config-reloader/0.log" Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.661416 1425 generic.go:296] "Generic (PLEG): container finished" podID=b59f4808-02df-4b2e-9694-8cf72a1b5e75 containerID="b9bc795189fd458b4ce14c6f192418f946a1c35d2f19ab6aedd4b4185f4bf487" exitCode=0 Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.661436 1425 generic.go:296] "Generic (PLEG): container finished" podID=b59f4808-02df-4b2e-9694-8cf72a1b5e75 containerID="6eba543bd270671b29706b350f4dfdf034230e4d33c811838d47249838c4ff2b" exitCode=0 Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.661452 1425 generic.go:296] "Generic (PLEG): container finished" podID=b59f4808-02df-4b2e-9694-8cf72a1b5e75 containerID="af5d20453c1aa7ea8d99d15e82767c98893d319f38bb9cd97aa815202e28d32c" exitCode=2 Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.661467 1425 generic.go:296] "Generic (PLEG): container finished" podID=b59f4808-02df-4b2e-9694-8cf72a1b5e75 containerID="e9efd1bb280b0736df20aa595c1c6f65c0d953e8a7a6a1dde105be0f728d6de6" exitCode=2 Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.661482 1425 generic.go:296] "Generic (PLEG): container finished" podID=b59f4808-02df-4b2e-9694-8cf72a1b5e75 containerID="902fedd514d362305b82903fec325b5adb139bd729665753e388c28b1e4d58bf" exitCode=0 Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.661522 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-1" event=&{ID:b59f4808-02df-4b2e-9694-8cf72a1b5e75 Type:ContainerDied Data:b9bc795189fd458b4ce14c6f192418f946a1c35d2f19ab6aedd4b4185f4bf487} Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.661549 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-1" event=&{ID:b59f4808-02df-4b2e-9694-8cf72a1b5e75 Type:ContainerDied Data:6eba543bd270671b29706b350f4dfdf034230e4d33c811838d47249838c4ff2b} Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.661568 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-1" event=&{ID:b59f4808-02df-4b2e-9694-8cf72a1b5e75 Type:ContainerDied Data:af5d20453c1aa7ea8d99d15e82767c98893d319f38bb9cd97aa815202e28d32c} Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.661583 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-1" event=&{ID:b59f4808-02df-4b2e-9694-8cf72a1b5e75 Type:ContainerDied Data:e9efd1bb280b0736df20aa595c1c6f65c0d953e8a7a6a1dde105be0f728d6de6} Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.661597 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-1" event=&{ID:b59f4808-02df-4b2e-9694-8cf72a1b5e75 Type:ContainerDied Data:902fedd514d362305b82903fec325b5adb139bd729665753e388c28b1e4d58bf} Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.663025 1425 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-1_8ac551e8-a959-4212-86a1-2b3ac482586f/prometheus-proxy/0.log" Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.663595 1425 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-1_8ac551e8-a959-4212-86a1-2b3ac482586f/config-reloader/0.log" Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.664076 1425 generic.go:296] "Generic (PLEG): container finished" podID=8ac551e8-a959-4212-86a1-2b3ac482586f containerID="e6208ed7b9782cddc95b22490d2dc4f820f1944cb39050a5d478fabf458cd570" exitCode=2 Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.664098 1425 generic.go:296] "Generic (PLEG): container finished" podID=8ac551e8-a959-4212-86a1-2b3ac482586f containerID="976ee1fd7cc0d02f224dbcf7a41285bac05287d0f5cf4e79c3f6c098c2a7dcd4" exitCode=0 Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.664113 1425 generic.go:296] "Generic (PLEG): container finished" podID=8ac551e8-a959-4212-86a1-2b3ac482586f containerID="254f6adadc261bc8b8bef366241b799f8c167b2d7c165cb79c16c15ee634f029" exitCode=2 Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.664128 1425 generic.go:296] "Generic (PLEG): container finished" podID=8ac551e8-a959-4212-86a1-2b3ac482586f containerID="6becde1d0c7f5d104d26d3a1c07efaa41bf091484c482dfffd15e87e4b5bf54c" exitCode=0 Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.664148 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-1" event=&{ID:8ac551e8-a959-4212-86a1-2b3ac482586f Type:ContainerDied Data:e6208ed7b9782cddc95b22490d2dc4f820f1944cb39050a5d478fabf458cd570} Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.664167 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-1" event=&{ID:8ac551e8-a959-4212-86a1-2b3ac482586f Type:ContainerDied Data:976ee1fd7cc0d02f224dbcf7a41285bac05287d0f5cf4e79c3f6c098c2a7dcd4} Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.664183 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-1" event=&{ID:8ac551e8-a959-4212-86a1-2b3ac482586f Type:ContainerDied Data:254f6adadc261bc8b8bef366241b799f8c167b2d7c165cb79c16c15ee634f029} Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.664198 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-1" event=&{ID:8ac551e8-a959-4212-86a1-2b3ac482586f Type:ContainerDied Data:6becde1d0c7f5d104d26d3a1c07efaa41bf091484c482dfffd15e87e4b5bf54c} Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.711748426Z" level=info msg="Removed container ff17947445f945fe096ba0ee1a903135c1548bf46b79e37353e7eddd5042f6f8: openshift-kube-storage-version-migrator/migrator-6cd87bdff8-p9dkf/migrator" id=8fcae286-e8c5-4e03-aa63-e13ff3deaf5f name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.712062 1425 scope.go:115] "RemoveContainer" containerID="ff17947445f945fe096ba0ee1a903135c1548bf46b79e37353e7eddd5042f6f8" Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:14:11.712626 1425 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff17947445f945fe096ba0ee1a903135c1548bf46b79e37353e7eddd5042f6f8\": container with ID starting with ff17947445f945fe096ba0ee1a903135c1548bf46b79e37353e7eddd5042f6f8 not found: ID does not exist" containerID="ff17947445f945fe096ba0ee1a903135c1548bf46b79e37353e7eddd5042f6f8" Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.712668 1425 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:ff17947445f945fe096ba0ee1a903135c1548bf46b79e37353e7eddd5042f6f8} err="failed to get container status \"ff17947445f945fe096ba0ee1a903135c1548bf46b79e37353e7eddd5042f6f8\": rpc error: code = NotFound desc = could not find container \"ff17947445f945fe096ba0ee1a903135c1548bf46b79e37353e7eddd5042f6f8\": container with ID starting with ff17947445f945fe096ba0ee1a903135c1548bf46b79e37353e7eddd5042f6f8 not found: ID does not exist" Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.780951 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjtff\" (UniqueName: \"kubernetes.io/projected/26f1eb05-de00-4fc2-864c-6b97e7f82e96-kube-api-access-hjtff\") pod \"26f1eb05-de00-4fc2-864c-6b97e7f82e96\" (UID: \"26f1eb05-de00-4fc2-864c-6b97e7f82e96\") " Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.868168 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26f1eb05-de00-4fc2-864c-6b97e7f82e96-kube-api-access-hjtff" (OuterVolumeSpecName: "kube-api-access-hjtff") pod "26f1eb05-de00-4fc2-864c-6b97e7f82e96" (UID: "26f1eb05-de00-4fc2-864c-6b97e7f82e96"). InnerVolumeSpecName "kube-api-access-hjtff". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.883628 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4df76\" (UniqueName: \"kubernetes.io/projected/89a48689-810c-464e-9ac6-d78977e60793-kube-api-access-4df76\") pod \"89a48689-810c-464e-9ac6-d78977e60793\" (UID: \"89a48689-810c-464e-9ac6-d78977e60793\") " Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.883832 1425 reconciler.go:399] "Volume detached for volume \"kube-api-access-hjtff\" (UniqueName: \"kubernetes.io/projected/26f1eb05-de00-4fc2-864c-6b97e7f82e96-kube-api-access-hjtff\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.921560 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89a48689-810c-464e-9ac6-d78977e60793-kube-api-access-4df76" (OuterVolumeSpecName: "kube-api-access-4df76") pod "89a48689-810c-464e-9ac6-d78977e60793" (UID: "89a48689-810c-464e-9ac6-d78977e60793"). InnerVolumeSpecName "kube-api-access-4df76". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: 2023-01-05T09:14:11Z [verbose] Del: openshift-monitoring:prometheus-operator-admission-webhook-598468bb49-54rjq:f8e5be3f-9ab7-484c-b61b-6231a8774e24:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.984271603Z" level=info msg="Stopped container 1a898502c172251a58f921f830b7d6d6f8f096259c3dca6353d228972a83e528: openshift-monitoring/prometheus-k8s-1/kube-rbac-proxy-thanos" id=38ba3ff4-e211-47bb-9c5c-b281cf8c00a6 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:11.985021 1425 reconciler.go:399] "Volume detached for volume \"kube-api-access-4df76\" (UniqueName: \"kubernetes.io/projected/89a48689-810c-464e-9ac6-d78977e60793-kube-api-access-4df76\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.986298207Z" level=info msg="Stopped container 67e4592536523011a4901ffad309428f68583ceb75cd44ec7d261c76a0778640: openshift-monitoring/prometheus-k8s-1/kube-rbac-proxy" id=ce076068-c46a-4e36-95a8-678dc15923c4 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.986812953Z" level=info msg="Stopping pod sandbox: bf998995ead8afb2642821b82636d7b867d8c0aa5ab164fe09e7396c2701f63b" id=443ae814-c81b-4828-85e3-65396d6f5257 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.987055764Z" level=info msg="Got pod network &{Name:prometheus-k8s-1 Namespace:openshift-monitoring ID:bf998995ead8afb2642821b82636d7b867d8c0aa5ab164fe09e7396c2701f63b UID:8ac551e8-a959-4212-86a1-2b3ac482586f NetNS:/var/run/netns/cf32a4f0-a7ea-4d0a-ada0-ddff6598e5f3 Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:14:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:11.987198581Z" level=info msg="Deleting pod openshift-monitoring_prometheus-k8s-1 from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.032064 1425 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-kube-storage-version-migrator/migrator-6cd87bdff8-p9dkf] Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.033252 1425 kubelet.go:2129] "SyncLoop REMOVE" source="api" pods=[openshift-kube-storage-version-migrator/migrator-6cd87bdff8-p9dkf] Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:12.047458872Z" level=info msg="Stopped container a30b14e5cf7f780de074100612e1db98e849c8b763e7b551ab0f6558754f69f6: openshift-monitoring/prometheus-adapter-5894cc86c7-47422/prometheus-adapter" id=7237ed66-7e91-4768-9e9c-1d49d92e0b75 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:12.047992255Z" level=info msg="Stopping pod sandbox: ef7044b787198dfb2aab7533e9c72c07b8c2ef842ecb652a788dc647c485157c" id=ac0da615-ba98-4ce0-b3b3-c6656349e4ac name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:12.048232636Z" level=info msg="Got pod network &{Name:prometheus-adapter-5894cc86c7-47422 Namespace:openshift-monitoring ID:ef7044b787198dfb2aab7533e9c72c07b8c2ef842ecb652a788dc647c485157c UID:a82fd83b-5344-4e7a-91dd-5894f3846e87 NetNS:/var/run/netns/5dddfea0-872a-4241-a44b-44a8fc6dedf6 Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:12.048378168Z" level=info msg="Deleting pod openshift-monitoring_prometheus-adapter-5894cc86c7-47422 from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:12.135535890Z" level=info msg="Stopped container 2bce691bc41bddf8da3b590cbdbb6b3b7604a9e34c167c7d2a14b15a944439f1: openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/kube-rbac-proxy-metrics" id=c087e6b1-94c4-4693-abec-7717fc3a4ba5 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:12.136048257Z" level=info msg="Stopping pod sandbox: 6c2cdef5be97661e24b16499c5f84baa5f0fdffbceb45f19b515c50e93bb6dd5" id=fa4dd68f-c048-41de-b9fe-b69465fa8d6d name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:12.136293341Z" level=info msg="Got pod network &{Name:thanos-querier-6f4f5fb7cf-wtcwx Namespace:openshift-monitoring ID:6c2cdef5be97661e24b16499c5f84baa5f0fdffbceb45f19b515c50e93bb6dd5 UID:72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c NetNS:/var/run/netns/0dfe86e0-4e47-48fc-98f3-38c9aac3ee85 Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:12.136468089Z" level=info msg="Deleting pod openshift-monitoring_thanos-querier-6f4f5fb7cf-wtcwx from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:12.192696033Z" level=info msg="Stopped pod sandbox: 8e21977f4f2aa77ecd7154452d4a15adc91d3b323c169dda297ad9c1f8901298" id=01f05c26-4ddf-4ed0-a1ec-c03c22a9e8d9 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:12.220921367Z" level=info msg="Stopped container 064809cddb0902a863e4fcd8d1da29b6a436aa3118d40b3b5ab6fbbda4fbf3d9: openshift-monitoring/alertmanager-main-1/kube-rbac-proxy-metric" id=377ca941-121b-46e2-8408-e511dc5fc965 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:12.221882288Z" level=info msg="Stopping pod sandbox: 4cc6b8c0009aed4f5500066da7f716df203dd1969281bb8dbce4d7052e55ec50" id=f5599b4c-21b4-4198-970a-636bd3f2d00b name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:12.222130071Z" level=info msg="Got pod network &{Name:alertmanager-main-1 Namespace:openshift-monitoring ID:4cc6b8c0009aed4f5500066da7f716df203dd1969281bb8dbce4d7052e55ec50 UID:b59f4808-02df-4b2e-9694-8cf72a1b5e75 NetNS:/var/run/netns/a22d0db9-a46e-42fb-931a-4c5793845fd1 Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:12.222276520Z" level=info msg="Deleting pod openshift-monitoring_alertmanager-main-1 from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: 2023-01-05T09:14:12Z [verbose] Del: openshift-monitoring:prometheus-k8s-1:8ac551e8-a959-4212-86a1-2b3ac482586f:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:12.367742421Z" level=info msg="Stopped pod sandbox: bf998995ead8afb2642821b82636d7b867d8c0aa5ab164fe09e7396c2701f63b" id=443ae814-c81b-4828-85e3-65396d6f5257 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.377205 1425 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-1_8ac551e8-a959-4212-86a1-2b3ac482586f/prometheus-proxy/0.log" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.377852 1425 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-1_8ac551e8-a959-4212-86a1-2b3ac482586f/config-reloader/0.log" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.413864 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/8ac551e8-a959-4212-86a1-2b3ac482586f-prometheus-k8s-db\") pod \"8ac551e8-a959-4212-86a1-2b3ac482586f\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.413917 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-grpc-tls\") pod \"8ac551e8-a959-4212-86a1-2b3ac482586f\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.413947 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-config\") pod \"8ac551e8-a959-4212-86a1-2b3ac482586f\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.413981 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-kube-etcd-client-certs\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-kube-etcd-client-certs\") pod \"8ac551e8-a959-4212-86a1-2b3ac482586f\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.414017 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"8ac551e8-a959-4212-86a1-2b3ac482586f\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.414048 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-metrics-client-ca\") pod \"8ac551e8-a959-4212-86a1-2b3ac482586f\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.414081 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8ac551e8-a959-4212-86a1-2b3ac482586f-tls-assets\") pod \"8ac551e8-a959-4212-86a1-2b3ac482586f\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.414115 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-configmap-metrics-client-ca\") pod \"8ac551e8-a959-4212-86a1-2b3ac482586f\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.414149 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-metrics-client-certs\") pod \"8ac551e8-a959-4212-86a1-2b3ac482586f\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.414183 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-prometheus-trusted-ca-bundle\") pod \"8ac551e8-a959-4212-86a1-2b3ac482586f\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.414216 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99gqh\" (UniqueName: \"kubernetes.io/projected/8ac551e8-a959-4212-86a1-2b3ac482586f-kube-api-access-99gqh\") pod \"8ac551e8-a959-4212-86a1-2b3ac482586f\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.414250 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-prometheus-k8s-rulefiles-0\") pod \"8ac551e8-a959-4212-86a1-2b3ac482586f\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.414283 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/f8e5be3f-9ab7-484c-b61b-6231a8774e24-tls-certificates\") pod \"f8e5be3f-9ab7-484c-b61b-6231a8774e24\" (UID: \"f8e5be3f-9ab7-484c-b61b-6231a8774e24\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.414315 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-configmap-kubelet-serving-ca-bundle\") pod \"8ac551e8-a959-4212-86a1-2b3ac482586f\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.414348 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-kube-rbac-proxy\") pod \"8ac551e8-a959-4212-86a1-2b3ac482586f\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.414381 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-configmap-serving-certs-ca-bundle\") pod \"8ac551e8-a959-4212-86a1-2b3ac482586f\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.414431 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-web-config\") pod \"8ac551e8-a959-4212-86a1-2b3ac482586f\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.414461 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-prometheus-k8s-tls\") pod \"8ac551e8-a959-4212-86a1-2b3ac482586f\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.414494 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-prometheus-k8s-proxy\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-prometheus-k8s-proxy\") pod \"8ac551e8-a959-4212-86a1-2b3ac482586f\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.414524 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8ac551e8-a959-4212-86a1-2b3ac482586f-config-out\") pod \"8ac551e8-a959-4212-86a1-2b3ac482586f\" (UID: \"8ac551e8-a959-4212-86a1-2b3ac482586f\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:14:12.414797 1425 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/8ac551e8-a959-4212-86a1-2b3ac482586f/volumes/kubernetes.io~empty-dir/config-out: clearQuota called, but quotas disabled Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.414951 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ac551e8-a959-4212-86a1-2b3ac482586f-config-out" (OuterVolumeSpecName: "config-out") pod "8ac551e8-a959-4212-86a1-2b3ac482586f" (UID: "8ac551e8-a959-4212-86a1-2b3ac482586f"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:14:12.415079 1425 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/8ac551e8-a959-4212-86a1-2b3ac482586f/volumes/kubernetes.io~empty-dir/prometheus-k8s-db: clearQuota called, but quotas disabled Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.416966 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ac551e8-a959-4212-86a1-2b3ac482586f-prometheus-k8s-db" (OuterVolumeSpecName: "prometheus-k8s-db") pod "8ac551e8-a959-4212-86a1-2b3ac482586f" (UID: "8ac551e8-a959-4212-86a1-2b3ac482586f"). InnerVolumeSpecName "prometheus-k8s-db". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:14:12.418124 1425 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/8ac551e8-a959-4212-86a1-2b3ac482586f/volumes/kubernetes.io~configmap/metrics-client-ca: clearQuota called, but quotas disabled Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.418313 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-metrics-client-ca" (OuterVolumeSpecName: "metrics-client-ca") pod "8ac551e8-a959-4212-86a1-2b3ac482586f" (UID: "8ac551e8-a959-4212-86a1-2b3ac482586f"). InnerVolumeSpecName "metrics-client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:14:12.418649 1425 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/8ac551e8-a959-4212-86a1-2b3ac482586f/volumes/kubernetes.io~configmap/configmap-metrics-client-ca: clearQuota called, but quotas disabled Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.418827 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-configmap-metrics-client-ca" (OuterVolumeSpecName: "configmap-metrics-client-ca") pod "8ac551e8-a959-4212-86a1-2b3ac482586f" (UID: "8ac551e8-a959-4212-86a1-2b3ac482586f"). InnerVolumeSpecName "configmap-metrics-client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:14:12.419130 1425 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/8ac551e8-a959-4212-86a1-2b3ac482586f/volumes/kubernetes.io~configmap/prometheus-trusted-ca-bundle: clearQuota called, but quotas disabled Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.419332 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-prometheus-trusted-ca-bundle" (OuterVolumeSpecName: "prometheus-trusted-ca-bundle") pod "8ac551e8-a959-4212-86a1-2b3ac482586f" (UID: "8ac551e8-a959-4212-86a1-2b3ac482586f"). InnerVolumeSpecName "prometheus-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:14:12.419681 1425 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/8ac551e8-a959-4212-86a1-2b3ac482586f/volumes/kubernetes.io~configmap/prometheus-k8s-rulefiles-0: clearQuota called, but quotas disabled Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.421724 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-prometheus-k8s-rulefiles-0" (OuterVolumeSpecName: "prometheus-k8s-rulefiles-0") pod "8ac551e8-a959-4212-86a1-2b3ac482586f" (UID: "8ac551e8-a959-4212-86a1-2b3ac482586f"). InnerVolumeSpecName "prometheus-k8s-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:14:12.422052 1425 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/8ac551e8-a959-4212-86a1-2b3ac482586f/volumes/kubernetes.io~configmap/configmap-kubelet-serving-ca-bundle: clearQuota called, but quotas disabled Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.422219 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-configmap-kubelet-serving-ca-bundle" (OuterVolumeSpecName: "configmap-kubelet-serving-ca-bundle") pod "8ac551e8-a959-4212-86a1-2b3ac482586f" (UID: "8ac551e8-a959-4212-86a1-2b3ac482586f"). InnerVolumeSpecName "configmap-kubelet-serving-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:14:12.422483 1425 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/8ac551e8-a959-4212-86a1-2b3ac482586f/volumes/kubernetes.io~configmap/configmap-serving-certs-ca-bundle: clearQuota called, but quotas disabled Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.422657 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-configmap-serving-certs-ca-bundle" (OuterVolumeSpecName: "configmap-serving-certs-ca-bundle") pod "8ac551e8-a959-4212-86a1-2b3ac482586f" (UID: "8ac551e8-a959-4212-86a1-2b3ac482586f"). InnerVolumeSpecName "configmap-serving-certs-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.450560 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-grpc-tls" (OuterVolumeSpecName: "secret-grpc-tls") pod "8ac551e8-a959-4212-86a1-2b3ac482586f" (UID: "8ac551e8-a959-4212-86a1-2b3ac482586f"). InnerVolumeSpecName "secret-grpc-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.450668 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ac551e8-a959-4212-86a1-2b3ac482586f-kube-api-access-99gqh" (OuterVolumeSpecName: "kube-api-access-99gqh") pod "8ac551e8-a959-4212-86a1-2b3ac482586f" (UID: "8ac551e8-a959-4212-86a1-2b3ac482586f"). InnerVolumeSpecName "kube-api-access-99gqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.457147 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-prometheus-k8s-thanos-sidecar-tls" (OuterVolumeSpecName: "secret-prometheus-k8s-thanos-sidecar-tls") pod "8ac551e8-a959-4212-86a1-2b3ac482586f" (UID: "8ac551e8-a959-4212-86a1-2b3ac482586f"). InnerVolumeSpecName "secret-prometheus-k8s-thanos-sidecar-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.457242 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-config" (OuterVolumeSpecName: "config") pod "8ac551e8-a959-4212-86a1-2b3ac482586f" (UID: "8ac551e8-a959-4212-86a1-2b3ac482586f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.457308 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-kube-etcd-client-certs" (OuterVolumeSpecName: "secret-kube-etcd-client-certs") pod "8ac551e8-a959-4212-86a1-2b3ac482586f" (UID: "8ac551e8-a959-4212-86a1-2b3ac482586f"). InnerVolumeSpecName "secret-kube-etcd-client-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.457406 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-metrics-client-certs" (OuterVolumeSpecName: "secret-metrics-client-certs") pod "8ac551e8-a959-4212-86a1-2b3ac482586f" (UID: "8ac551e8-a959-4212-86a1-2b3ac482586f"). InnerVolumeSpecName "secret-metrics-client-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.457478 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-kube-rbac-proxy" (OuterVolumeSpecName: "secret-kube-rbac-proxy") pod "8ac551e8-a959-4212-86a1-2b3ac482586f" (UID: "8ac551e8-a959-4212-86a1-2b3ac482586f"). InnerVolumeSpecName "secret-kube-rbac-proxy". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.458825 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ac551e8-a959-4212-86a1-2b3ac482586f-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "8ac551e8-a959-4212-86a1-2b3ac482586f" (UID: "8ac551e8-a959-4212-86a1-2b3ac482586f"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: 2023-01-05T09:14:12Z [verbose] Del: openshift-monitoring:prometheus-adapter-5894cc86c7-47422:a82fd83b-5344-4e7a-91dd-5894f3846e87:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.470466 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8e5be3f-9ab7-484c-b61b-6231a8774e24-tls-certificates" (OuterVolumeSpecName: "tls-certificates") pod "f8e5be3f-9ab7-484c-b61b-6231a8774e24" (UID: "f8e5be3f-9ab7-484c-b61b-6231a8774e24"). InnerVolumeSpecName "tls-certificates". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.487152 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-prometheus-k8s-tls" (OuterVolumeSpecName: "secret-prometheus-k8s-tls") pod "8ac551e8-a959-4212-86a1-2b3ac482586f" (UID: "8ac551e8-a959-4212-86a1-2b3ac482586f"). InnerVolumeSpecName "secret-prometheus-k8s-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.492061 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-prometheus-k8s-proxy" (OuterVolumeSpecName: "secret-prometheus-k8s-proxy") pod "8ac551e8-a959-4212-86a1-2b3ac482586f" (UID: "8ac551e8-a959-4212-86a1-2b3ac482586f"). InnerVolumeSpecName "secret-prometheus-k8s-proxy". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.495016 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-web-config" (OuterVolumeSpecName: "web-config") pod "8ac551e8-a959-4212-86a1-2b3ac482586f" (UID: "8ac551e8-a959-4212-86a1-2b3ac482586f"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.514913 1425 reconciler.go:399] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-config\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.514958 1425 reconciler.go:399] "Volume detached for volume \"secret-kube-etcd-client-certs\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-kube-etcd-client-certs\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.514979 1425 reconciler.go:399] "Volume detached for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-prometheus-k8s-thanos-sidecar-tls\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.514998 1425 reconciler.go:399] "Volume detached for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-metrics-client-ca\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.515015 1425 reconciler.go:399] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8ac551e8-a959-4212-86a1-2b3ac482586f-tls-assets\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.515033 1425 reconciler.go:399] "Volume detached for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-configmap-metrics-client-ca\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.515066 1425 reconciler.go:399] "Volume detached for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-metrics-client-certs\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.515084 1425 reconciler.go:399] "Volume detached for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-prometheus-trusted-ca-bundle\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.515101 1425 reconciler.go:399] "Volume detached for volume \"kube-api-access-99gqh\" (UniqueName: \"kubernetes.io/projected/8ac551e8-a959-4212-86a1-2b3ac482586f-kube-api-access-99gqh\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.515119 1425 reconciler.go:399] "Volume detached for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-prometheus-k8s-rulefiles-0\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.515138 1425 reconciler.go:399] "Volume detached for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/f8e5be3f-9ab7-484c-b61b-6231a8774e24-tls-certificates\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.515157 1425 reconciler.go:399] "Volume detached for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-configmap-kubelet-serving-ca-bundle\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.515175 1425 reconciler.go:399] "Volume detached for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-kube-rbac-proxy\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.515194 1425 reconciler.go:399] "Volume detached for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8ac551e8-a959-4212-86a1-2b3ac482586f-configmap-serving-certs-ca-bundle\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.515210 1425 reconciler.go:399] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-web-config\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.515230 1425 reconciler.go:399] "Volume detached for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-prometheus-k8s-tls\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.515248 1425 reconciler.go:399] "Volume detached for volume \"secret-prometheus-k8s-proxy\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-prometheus-k8s-proxy\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.515264 1425 reconciler.go:399] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8ac551e8-a959-4212-86a1-2b3ac482586f-config-out\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.515281 1425 reconciler.go:399] "Volume detached for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/8ac551e8-a959-4212-86a1-2b3ac482586f-prometheus-k8s-db\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.515299 1425 reconciler.go:399] "Volume detached for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/8ac551e8-a959-4212-86a1-2b3ac482586f-secret-grpc-tls\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:12.553729224Z" level=info msg="Stopped pod sandbox: ef7044b787198dfb2aab7533e9c72c07b8c2ef842ecb652a788dc647c485157c" id=ac0da615-ba98-4ce0-b3b3-c6656349e4ac name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.615601 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a82fd83b-5344-4e7a-91dd-5894f3846e87-serving-certs-ca-bundle\") pod \"a82fd83b-5344-4e7a-91dd-5894f3846e87\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.615662 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"prometheus-adapter-prometheus-config\" (UniqueName: \"kubernetes.io/configmap/a82fd83b-5344-4e7a-91dd-5894f3846e87-prometheus-adapter-prometheus-config\") pod \"a82fd83b-5344-4e7a-91dd-5894f3846e87\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.615698 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a82fd83b-5344-4e7a-91dd-5894f3846e87-tmpfs\") pod \"a82fd83b-5344-4e7a-91dd-5894f3846e87\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.615737 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjb9d\" (UniqueName: \"kubernetes.io/projected/a82fd83b-5344-4e7a-91dd-5894f3846e87-kube-api-access-kjb9d\") pod \"a82fd83b-5344-4e7a-91dd-5894f3846e87\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.615775 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"prometheus-adapter-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/a82fd83b-5344-4e7a-91dd-5894f3846e87-prometheus-adapter-audit-profiles\") pod \"a82fd83b-5344-4e7a-91dd-5894f3846e87\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.615812 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"tls\" (UniqueName: \"kubernetes.io/secret/a82fd83b-5344-4e7a-91dd-5894f3846e87-tls\") pod \"a82fd83b-5344-4e7a-91dd-5894f3846e87\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.615847 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a82fd83b-5344-4e7a-91dd-5894f3846e87-config\") pod \"a82fd83b-5344-4e7a-91dd-5894f3846e87\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.615878 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/a82fd83b-5344-4e7a-91dd-5894f3846e87-audit-log\") pod \"a82fd83b-5344-4e7a-91dd-5894f3846e87\" (UID: \"a82fd83b-5344-4e7a-91dd-5894f3846e87\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:14:12.616123 1425 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/a82fd83b-5344-4e7a-91dd-5894f3846e87/volumes/kubernetes.io~empty-dir/audit-log: clearQuota called, but quotas disabled Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.616251 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a82fd83b-5344-4e7a-91dd-5894f3846e87-audit-log" (OuterVolumeSpecName: "audit-log") pod "a82fd83b-5344-4e7a-91dd-5894f3846e87" (UID: "a82fd83b-5344-4e7a-91dd-5894f3846e87"). InnerVolumeSpecName "audit-log". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:14:12.616410 1425 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/a82fd83b-5344-4e7a-91dd-5894f3846e87/volumes/kubernetes.io~configmap/serving-certs-ca-bundle: clearQuota called, but quotas disabled Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.616617 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a82fd83b-5344-4e7a-91dd-5894f3846e87-serving-certs-ca-bundle" (OuterVolumeSpecName: "serving-certs-ca-bundle") pod "a82fd83b-5344-4e7a-91dd-5894f3846e87" (UID: "a82fd83b-5344-4e7a-91dd-5894f3846e87"). InnerVolumeSpecName "serving-certs-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:14:12.616742 1425 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/a82fd83b-5344-4e7a-91dd-5894f3846e87/volumes/kubernetes.io~configmap/prometheus-adapter-prometheus-config: clearQuota called, but quotas disabled Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.616920 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a82fd83b-5344-4e7a-91dd-5894f3846e87-prometheus-adapter-prometheus-config" (OuterVolumeSpecName: "prometheus-adapter-prometheus-config") pod "a82fd83b-5344-4e7a-91dd-5894f3846e87" (UID: "a82fd83b-5344-4e7a-91dd-5894f3846e87"). InnerVolumeSpecName "prometheus-adapter-prometheus-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:14:12.617029 1425 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/a82fd83b-5344-4e7a-91dd-5894f3846e87/volumes/kubernetes.io~empty-dir/tmpfs: clearQuota called, but quotas disabled Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.617062 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a82fd83b-5344-4e7a-91dd-5894f3846e87-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "a82fd83b-5344-4e7a-91dd-5894f3846e87" (UID: "a82fd83b-5344-4e7a-91dd-5894f3846e87"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:14:12.618767 1425 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/a82fd83b-5344-4e7a-91dd-5894f3846e87/volumes/kubernetes.io~configmap/prometheus-adapter-audit-profiles: clearQuota called, but quotas disabled Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.619103 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a82fd83b-5344-4e7a-91dd-5894f3846e87-prometheus-adapter-audit-profiles" (OuterVolumeSpecName: "prometheus-adapter-audit-profiles") pod "a82fd83b-5344-4e7a-91dd-5894f3846e87" (UID: "a82fd83b-5344-4e7a-91dd-5894f3846e87"). InnerVolumeSpecName "prometheus-adapter-audit-profiles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:14:12.619265 1425 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/a82fd83b-5344-4e7a-91dd-5894f3846e87/volumes/kubernetes.io~configmap/config: clearQuota called, but quotas disabled Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.619461 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a82fd83b-5344-4e7a-91dd-5894f3846e87-config" (OuterVolumeSpecName: "config") pod "a82fd83b-5344-4e7a-91dd-5894f3846e87" (UID: "a82fd83b-5344-4e7a-91dd-5894f3846e87"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.633132 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a82fd83b-5344-4e7a-91dd-5894f3846e87-tls" (OuterVolumeSpecName: "tls") pod "a82fd83b-5344-4e7a-91dd-5894f3846e87" (UID: "a82fd83b-5344-4e7a-91dd-5894f3846e87"). InnerVolumeSpecName "tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.633239 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a82fd83b-5344-4e7a-91dd-5894f3846e87-kube-api-access-kjb9d" (OuterVolumeSpecName: "kube-api-access-kjb9d") pod "a82fd83b-5344-4e7a-91dd-5894f3846e87" (UID: "a82fd83b-5344-4e7a-91dd-5894f3846e87"). InnerVolumeSpecName "kube-api-access-kjb9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: 2023-01-05T09:14:12Z [verbose] Del: openshift-monitoring:thanos-querier-6f4f5fb7cf-wtcwx:72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.674007 1425 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=26f1eb05-de00-4fc2-864c-6b97e7f82e96 path="/var/lib/kubelet/pods/26f1eb05-de00-4fc2-864c-6b97e7f82e96/volumes" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.674714 1425 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-1_8ac551e8-a959-4212-86a1-2b3ac482586f/prometheus-proxy/0.log" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.713024 1425 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-1_8ac551e8-a959-4212-86a1-2b3ac482586f/config-reloader/0.log" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.714086 1425 generic.go:296] "Generic (PLEG): container finished" podID=8ac551e8-a959-4212-86a1-2b3ac482586f containerID="1a898502c172251a58f921f830b7d6d6f8f096259c3dca6353d228972a83e528" exitCode=0 Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.714109 1425 generic.go:296] "Generic (PLEG): container finished" podID=8ac551e8-a959-4212-86a1-2b3ac482586f containerID="67e4592536523011a4901ffad309428f68583ceb75cd44ec7d261c76a0778640" exitCode=0 Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.714159 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-1" event=&{ID:8ac551e8-a959-4212-86a1-2b3ac482586f Type:ContainerDied Data:1a898502c172251a58f921f830b7d6d6f8f096259c3dca6353d228972a83e528} Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.714186 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-1" event=&{ID:8ac551e8-a959-4212-86a1-2b3ac482586f Type:ContainerDied Data:67e4592536523011a4901ffad309428f68583ceb75cd44ec7d261c76a0778640} Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.714203 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-1" event=&{ID:8ac551e8-a959-4212-86a1-2b3ac482586f Type:ContainerDied Data:bf998995ead8afb2642821b82636d7b867d8c0aa5ab164fe09e7396c2701f63b} Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.714223 1425 scope.go:115] "RemoveContainer" containerID="1a898502c172251a58f921f830b7d6d6f8f096259c3dca6353d228972a83e528" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.719192 1425 reconciler.go:399] "Volume detached for volume \"kube-api-access-kjb9d\" (UniqueName: \"kubernetes.io/projected/a82fd83b-5344-4e7a-91dd-5894f3846e87-kube-api-access-kjb9d\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.719217 1425 reconciler.go:399] "Volume detached for volume \"prometheus-adapter-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/a82fd83b-5344-4e7a-91dd-5894f3846e87-prometheus-adapter-audit-profiles\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.719233 1425 reconciler.go:399] "Volume detached for volume \"tls\" (UniqueName: \"kubernetes.io/secret/a82fd83b-5344-4e7a-91dd-5894f3846e87-tls\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.719249 1425 reconciler.go:399] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a82fd83b-5344-4e7a-91dd-5894f3846e87-config\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.719263 1425 reconciler.go:399] "Volume detached for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/a82fd83b-5344-4e7a-91dd-5894f3846e87-audit-log\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.719279 1425 reconciler.go:399] "Volume detached for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a82fd83b-5344-4e7a-91dd-5894f3846e87-serving-certs-ca-bundle\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.719299 1425 reconciler.go:399] "Volume detached for volume \"prometheus-adapter-prometheus-config\" (UniqueName: \"kubernetes.io/configmap/a82fd83b-5344-4e7a-91dd-5894f3846e87-prometheus-adapter-prometheus-config\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.719315 1425 reconciler.go:399] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a82fd83b-5344-4e7a-91dd-5894f3846e87-tmpfs\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:12.727651696Z" level=info msg="Removing container: 1a898502c172251a58f921f830b7d6d6f8f096259c3dca6353d228972a83e528" id=c8eb814f-8725-4f6d-b978-128f6e34e37b name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:12.738573626Z" level=info msg="Stopped pod sandbox: 6c2cdef5be97661e24b16499c5f84baa5f0fdffbceb45f19b515c50e93bb6dd5" id=fa4dd68f-c048-41de-b9fe-b69465fa8d6d name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.743967 1425 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-6f4f5fb7cf-wtcwx_72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c/oauth-proxy/0.log" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.745069 1425 generic.go:296] "Generic (PLEG): container finished" podID=72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c containerID="2bce691bc41bddf8da3b590cbdbb6b3b7604a9e34c167c7d2a14b15a944439f1" exitCode=0 Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.745125 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" event=&{ID:72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c Type:ContainerDied Data:2bce691bc41bddf8da3b590cbdbb6b3b7604a9e34c167c7d2a14b15a944439f1} Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.747259 1425 generic.go:296] "Generic (PLEG): container finished" podID=a82fd83b-5344-4e7a-91dd-5894f3846e87 containerID="a30b14e5cf7f780de074100612e1db98e849c8b763e7b551ab0f6558754f69f6" exitCode=0 Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.747314 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" event=&{ID:a82fd83b-5344-4e7a-91dd-5894f3846e87 Type:ContainerDied Data:a30b14e5cf7f780de074100612e1db98e849c8b763e7b551ab0f6558754f69f6} Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.747336 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-47422" event=&{ID:a82fd83b-5344-4e7a-91dd-5894f3846e87 Type:ContainerDied Data:ef7044b787198dfb2aab7533e9c72c07b8c2ef842ecb652a788dc647c485157c} Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.766874 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-54rjq" event=&{ID:f8e5be3f-9ab7-484c-b61b-6231a8774e24 Type:ContainerDied Data:8e21977f4f2aa77ecd7154452d4a15adc91d3b323c169dda297ad9c1f8901298} Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.771192 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-6775957967-zsr9c" event=&{ID:89a48689-810c-464e-9ac6-d78977e60793 Type:ContainerDied Data:e7532666db14128bc848af4cbb52adc603db1ec585b807063b42799df403a10b} Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.776430 1425 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-1_b59f4808-02df-4b2e-9694-8cf72a1b5e75/alertmanager-proxy/0.log" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.776843 1425 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-1_b59f4808-02df-4b2e-9694-8cf72a1b5e75/config-reloader/0.log" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.777231 1425 generic.go:296] "Generic (PLEG): container finished" podID=b59f4808-02df-4b2e-9694-8cf72a1b5e75 containerID="064809cddb0902a863e4fcd8d1da29b6a436aa3118d40b3b5ab6fbbda4fbf3d9" exitCode=0 Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.777259 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-1" event=&{ID:b59f4808-02df-4b2e-9694-8cf72a1b5e75 Type:ContainerDied Data:064809cddb0902a863e4fcd8d1da29b6a436aa3118d40b3b5ab6fbbda4fbf3d9} Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.789079 1425 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-6f4f5fb7cf-wtcwx_72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c/oauth-proxy/0.log" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.853973 1425 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-monitoring/prometheus-k8s-1] Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.854025 1425 kubelet.go:2129] "SyncLoop REMOVE" source="api" pods=[openshift-monitoring/prometheus-k8s-1] Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.906955 1425 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-monitoring/prometheus-adapter-5894cc86c7-47422] Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:12.909279595Z" level=info msg="Removed container 1a898502c172251a58f921f830b7d6d6f8f096259c3dca6353d228972a83e528: openshift-monitoring/prometheus-k8s-1/kube-rbac-proxy-thanos" id=c8eb814f-8725-4f6d-b978-128f6e34e37b name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.910086 1425 scope.go:115] "RemoveContainer" containerID="67e4592536523011a4901ffad309428f68583ceb75cd44ec7d261c76a0778640" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:12.911244303Z" level=info msg="Removing container: 67e4592536523011a4901ffad309428f68583ceb75cd44ec7d261c76a0778640" id=65948306-4e31-47a5-b5dc-d20354d8b92f name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.928835 1425 kubelet.go:2129] "SyncLoop REMOVE" source="api" pods=[openshift-monitoring/prometheus-adapter-5894cc86c7-47422] Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.958721 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-metrics-client-ca\") pod \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.958774 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-thanos-querier-tls\") pod \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.958819 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"thanos-querier-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-thanos-querier-trusted-ca-bundle\") pod \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.958851 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-thanos-querier-oauth-cookie\" (UniqueName: \"kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-thanos-querier-oauth-cookie\") pod \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.958888 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.958921 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brfpd\" (UniqueName: \"kubernetes.io/projected/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-kube-api-access-brfpd\") pod \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.958954 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.958984 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-grpc-tls\") pod \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.959017 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-thanos-querier-kube-rbac-proxy\") pod \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\" (UID: \"72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c\") " Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.967174 1425 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-54rjq] Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:14:12.967620 1425 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c/volumes/kubernetes.io~configmap/metrics-client-ca: clearQuota called, but quotas disabled Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.967835 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-metrics-client-ca" (OuterVolumeSpecName: "metrics-client-ca") pod "72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c" (UID: "72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c"). InnerVolumeSpecName "metrics-client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:14:12.968202 1425 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c/volumes/kubernetes.io~configmap/thanos-querier-trusted-ca-bundle: clearQuota called, but quotas disabled Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.968455 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-thanos-querier-trusted-ca-bundle" (OuterVolumeSpecName: "thanos-querier-trusted-ca-bundle") pod "72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c" (UID: "72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c"). InnerVolumeSpecName "thanos-querier-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.974141 1425 kubelet.go:2129] "SyncLoop REMOVE" source="api" pods=[openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-54rjq] Jan 05 09:14:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:12.985714 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-thanos-querier-kube-rbac-proxy" (OuterVolumeSpecName: "secret-thanos-querier-kube-rbac-proxy") pod "72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c" (UID: "72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c"). InnerVolumeSpecName "secret-thanos-querier-kube-rbac-proxy". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.023878 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-thanos-querier-oauth-cookie" (OuterVolumeSpecName: "secret-thanos-querier-oauth-cookie") pod "72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c" (UID: "72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c"). InnerVolumeSpecName "secret-thanos-querier-oauth-cookie". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.025171 1425 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-network-diagnostics/network-check-source-6775957967-zsr9c] Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.032761 1425 kubelet.go:2129] "SyncLoop REMOVE" source="api" pods=[openshift-network-diagnostics/network-check-source-6775957967-zsr9c] Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.036025 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-thanos-querier-tls" (OuterVolumeSpecName: "secret-thanos-querier-tls") pod "72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c" (UID: "72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c"). InnerVolumeSpecName "secret-thanos-querier-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.036128 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-kube-api-access-brfpd" (OuterVolumeSpecName: "kube-api-access-brfpd") pod "72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c" (UID: "72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c"). InnerVolumeSpecName "kube-api-access-brfpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: 2023-01-05T09:14:12Z [verbose] Del: openshift-monitoring:alertmanager-main-1:b59f4808-02df-4b2e-9694-8cf72a1b5e75:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.047304 1425 status_manager.go:652] "Status for pod is up-to-date; skipping" podUID=8ac551e8-a959-4212-86a1-2b3ac482586f Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.071668 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-thanos-querier-kube-rbac-proxy-metrics" (OuterVolumeSpecName: "secret-thanos-querier-kube-rbac-proxy-metrics") pod "72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c" (UID: "72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c"). InnerVolumeSpecName "secret-thanos-querier-kube-rbac-proxy-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.071676 1425 reconciler.go:399] "Volume detached for volume \"kube-api-access-brfpd\" (UniqueName: \"kubernetes.io/projected/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-kube-api-access-brfpd\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.071840 1425 reconciler.go:399] "Volume detached for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-thanos-querier-kube-rbac-proxy\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.071863 1425 reconciler.go:399] "Volume detached for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-metrics-client-ca\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.071881 1425 reconciler.go:399] "Volume detached for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-thanos-querier-tls\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.071900 1425 reconciler.go:399] "Volume detached for volume \"secret-thanos-querier-oauth-cookie\" (UniqueName: \"kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-thanos-querier-oauth-cookie\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.071918 1425 reconciler.go:399] "Volume detached for volume \"thanos-querier-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-thanos-querier-trusted-ca-bundle\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.072114 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-thanos-querier-kube-rbac-proxy-rules" (OuterVolumeSpecName: "secret-thanos-querier-kube-rbac-proxy-rules") pod "72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c" (UID: "72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c"). InnerVolumeSpecName "secret-thanos-querier-kube-rbac-proxy-rules". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:13.072887477Z" level=info msg="Removed container 67e4592536523011a4901ffad309428f68583ceb75cd44ec7d261c76a0778640: openshift-monitoring/prometheus-k8s-1/kube-rbac-proxy" id=65948306-4e31-47a5-b5dc-d20354d8b92f name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.073258 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-grpc-tls" (OuterVolumeSpecName: "secret-grpc-tls") pod "72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c" (UID: "72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c"). InnerVolumeSpecName "secret-grpc-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.079056 1425 scope.go:115] "RemoveContainer" containerID="e6208ed7b9782cddc95b22490d2dc4f820f1944cb39050a5d478fabf458cd570" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:13.081330722Z" level=info msg="Removing container: e6208ed7b9782cddc95b22490d2dc4f820f1944cb39050a5d478fabf458cd570" id=b17dbf6a-3058-4290-945a-fb20542771cf name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:13.117202089Z" level=info msg="Removed container e6208ed7b9782cddc95b22490d2dc4f820f1944cb39050a5d478fabf458cd570: openshift-monitoring/prometheus-k8s-1/prometheus-proxy" id=b17dbf6a-3058-4290-945a-fb20542771cf name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.117502 1425 scope.go:115] "RemoveContainer" containerID="976ee1fd7cc0d02f224dbcf7a41285bac05287d0f5cf4e79c3f6c098c2a7dcd4" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:13.119817946Z" level=info msg="Removing container: 976ee1fd7cc0d02f224dbcf7a41285bac05287d0f5cf4e79c3f6c098c2a7dcd4" id=031fc5c9-fdd5-4616-b678-376dd73c6f90 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:13.148747427Z" level=info msg="Stopped pod sandbox: 4cc6b8c0009aed4f5500066da7f716df203dd1969281bb8dbce4d7052e55ec50" id=f5599b4c-21b4-4198-970a-636bd3f2d00b name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:13.160187374Z" level=info msg="Removed container 976ee1fd7cc0d02f224dbcf7a41285bac05287d0f5cf4e79c3f6c098c2a7dcd4: openshift-monitoring/prometheus-k8s-1/thanos-sidecar" id=031fc5c9-fdd5-4616-b678-376dd73c6f90 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.161728 1425 scope.go:115] "RemoveContainer" containerID="254f6adadc261bc8b8bef366241b799f8c167b2d7c165cb79c16c15ee634f029" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.178815 1425 reconciler.go:399] "Volume detached for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-thanos-querier-kube-rbac-proxy-metrics\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.178853 1425 reconciler.go:399] "Volume detached for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-thanos-querier-kube-rbac-proxy-rules\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.178873 1425 reconciler.go:399] "Volume detached for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c-secret-grpc-tls\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:13.192378247Z" level=info msg="Removing container: 254f6adadc261bc8b8bef366241b799f8c167b2d7c165cb79c16c15ee634f029" id=e808a3d6-a62f-461f-be71-23cd2fa4a0f0 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.203446 1425 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-1_b59f4808-02df-4b2e-9694-8cf72a1b5e75/alertmanager-proxy/0.log" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.204471 1425 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-1_b59f4808-02df-4b2e-9694-8cf72a1b5e75/config-reloader/0.log" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:13.258492047Z" level=info msg="Removed container 254f6adadc261bc8b8bef366241b799f8c167b2d7c165cb79c16c15ee634f029: openshift-monitoring/prometheus-k8s-1/config-reloader" id=e808a3d6-a62f-461f-be71-23cd2fa4a0f0 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.259041 1425 scope.go:115] "RemoveContainer" containerID="6becde1d0c7f5d104d26d3a1c07efaa41bf091484c482dfffd15e87e4b5bf54c" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:13.260025721Z" level=info msg="Removing container: 6becde1d0c7f5d104d26d3a1c07efaa41bf091484c482dfffd15e87e4b5bf54c" id=063ef4d8-862f-4b5d-8cbf-1ffa6f036f31 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:13.354043388Z" level=info msg="Removed container 6becde1d0c7f5d104d26d3a1c07efaa41bf091484c482dfffd15e87e4b5bf54c: openshift-monitoring/prometheus-k8s-1/prometheus" id=063ef4d8-862f-4b5d-8cbf-1ffa6f036f31 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.354475 1425 scope.go:115] "RemoveContainer" containerID="c2fc2489b690ab18d10fc72750d26008cb07a65be8420998a29168a6a7cea2af" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:13.355673043Z" level=info msg="Removing container: c2fc2489b690ab18d10fc72750d26008cb07a65be8420998a29168a6a7cea2af" id=a2cfffbc-892f-4a4e-b85c-b7df4145ae4d name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.381050 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-web-config\") pod \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.381103 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-alertmanager-main-proxy\" (UniqueName: \"kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-secret-alertmanager-main-proxy\") pod \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.381136 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-config-volume\") pod \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.381170 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/b59f4808-02df-4b2e-9694-8cf72a1b5e75-metrics-client-ca\") pod \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.381206 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-secret-alertmanager-main-tls\") pod \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.381241 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/b59f4808-02df-4b2e-9694-8cf72a1b5e75-alertmanager-main-db\") pod \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.381275 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b59f4808-02df-4b2e-9694-8cf72a1b5e75-config-out\") pod \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.381311 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqgp9\" (UniqueName: \"kubernetes.io/projected/b59f4808-02df-4b2e-9694-8cf72a1b5e75-kube-api-access-xqgp9\") pod \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.381348 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-secret-alertmanager-kube-rbac-proxy\") pod \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.381407 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b59f4808-02df-4b2e-9694-8cf72a1b5e75-tls-assets\") pod \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.381442 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b59f4808-02df-4b2e-9694-8cf72a1b5e75-alertmanager-trusted-ca-bundle\") pod \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.381479 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-secret-alertmanager-kube-rbac-proxy-metric\") pod \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\" (UID: \"b59f4808-02df-4b2e-9694-8cf72a1b5e75\") " Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:14:13.381998 1425 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/b59f4808-02df-4b2e-9694-8cf72a1b5e75/volumes/kubernetes.io~empty-dir/alertmanager-main-db: clearQuota called, but quotas disabled Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.382104 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b59f4808-02df-4b2e-9694-8cf72a1b5e75-alertmanager-main-db" (OuterVolumeSpecName: "alertmanager-main-db") pod "b59f4808-02df-4b2e-9694-8cf72a1b5e75" (UID: "b59f4808-02df-4b2e-9694-8cf72a1b5e75"). InnerVolumeSpecName "alertmanager-main-db". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:14:13.382858 1425 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/b59f4808-02df-4b2e-9694-8cf72a1b5e75/volumes/kubernetes.io~configmap/metrics-client-ca: clearQuota called, but quotas disabled Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.383068 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b59f4808-02df-4b2e-9694-8cf72a1b5e75-metrics-client-ca" (OuterVolumeSpecName: "metrics-client-ca") pod "b59f4808-02df-4b2e-9694-8cf72a1b5e75" (UID: "b59f4808-02df-4b2e-9694-8cf72a1b5e75"). InnerVolumeSpecName "metrics-client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:14:13.383625 1425 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/b59f4808-02df-4b2e-9694-8cf72a1b5e75/volumes/kubernetes.io~empty-dir/config-out: clearQuota called, but quotas disabled Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.383742 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b59f4808-02df-4b2e-9694-8cf72a1b5e75-config-out" (OuterVolumeSpecName: "config-out") pod "b59f4808-02df-4b2e-9694-8cf72a1b5e75" (UID: "b59f4808-02df-4b2e-9694-8cf72a1b5e75"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:14:13.384247 1425 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/b59f4808-02df-4b2e-9694-8cf72a1b5e75/volumes/kubernetes.io~configmap/alertmanager-trusted-ca-bundle: clearQuota called, but quotas disabled Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.384479 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b59f4808-02df-4b2e-9694-8cf72a1b5e75-alertmanager-trusted-ca-bundle" (OuterVolumeSpecName: "alertmanager-trusted-ca-bundle") pod "b59f4808-02df-4b2e-9694-8cf72a1b5e75" (UID: "b59f4808-02df-4b2e-9694-8cf72a1b5e75"). InnerVolumeSpecName "alertmanager-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.390990 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-secret-alertmanager-kube-rbac-proxy-metric" (OuterVolumeSpecName: "secret-alertmanager-kube-rbac-proxy-metric") pod "b59f4808-02df-4b2e-9694-8cf72a1b5e75" (UID: "b59f4808-02df-4b2e-9694-8cf72a1b5e75"). InnerVolumeSpecName "secret-alertmanager-kube-rbac-proxy-metric". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.400817 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-secret-alertmanager-main-proxy" (OuterVolumeSpecName: "secret-alertmanager-main-proxy") pod "b59f4808-02df-4b2e-9694-8cf72a1b5e75" (UID: "b59f4808-02df-4b2e-9694-8cf72a1b5e75"). InnerVolumeSpecName "secret-alertmanager-main-proxy". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.400886 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-config-volume" (OuterVolumeSpecName: "config-volume") pod "b59f4808-02df-4b2e-9694-8cf72a1b5e75" (UID: "b59f4808-02df-4b2e-9694-8cf72a1b5e75"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.405695 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-secret-alertmanager-kube-rbac-proxy" (OuterVolumeSpecName: "secret-alertmanager-kube-rbac-proxy") pod "b59f4808-02df-4b2e-9694-8cf72a1b5e75" (UID: "b59f4808-02df-4b2e-9694-8cf72a1b5e75"). InnerVolumeSpecName "secret-alertmanager-kube-rbac-proxy". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.405768 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-secret-alertmanager-main-tls" (OuterVolumeSpecName: "secret-alertmanager-main-tls") pod "b59f4808-02df-4b2e-9694-8cf72a1b5e75" (UID: "b59f4808-02df-4b2e-9694-8cf72a1b5e75"). InnerVolumeSpecName "secret-alertmanager-main-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.406915 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b59f4808-02df-4b2e-9694-8cf72a1b5e75-kube-api-access-xqgp9" (OuterVolumeSpecName: "kube-api-access-xqgp9") pod "b59f4808-02df-4b2e-9694-8cf72a1b5e75" (UID: "b59f4808-02df-4b2e-9694-8cf72a1b5e75"). InnerVolumeSpecName "kube-api-access-xqgp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:13.407609191Z" level=info msg="Removed container c2fc2489b690ab18d10fc72750d26008cb07a65be8420998a29168a6a7cea2af: openshift-monitoring/prometheus-k8s-1/init-config-reloader" id=a2cfffbc-892f-4a4e-b85c-b7df4145ae4d name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.408069 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b59f4808-02df-4b2e-9694-8cf72a1b5e75-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "b59f4808-02df-4b2e-9694-8cf72a1b5e75" (UID: "b59f4808-02df-4b2e-9694-8cf72a1b5e75"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.408167 1425 scope.go:115] "RemoveContainer" containerID="1a898502c172251a58f921f830b7d6d6f8f096259c3dca6353d228972a83e528" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.408334 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-web-config" (OuterVolumeSpecName: "web-config") pod "b59f4808-02df-4b2e-9694-8cf72a1b5e75" (UID: "b59f4808-02df-4b2e-9694-8cf72a1b5e75"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:14:13.408624 1425 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a898502c172251a58f921f830b7d6d6f8f096259c3dca6353d228972a83e528\": container with ID starting with 1a898502c172251a58f921f830b7d6d6f8f096259c3dca6353d228972a83e528 not found: ID does not exist" containerID="1a898502c172251a58f921f830b7d6d6f8f096259c3dca6353d228972a83e528" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.408659 1425 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:1a898502c172251a58f921f830b7d6d6f8f096259c3dca6353d228972a83e528} err="failed to get container status \"1a898502c172251a58f921f830b7d6d6f8f096259c3dca6353d228972a83e528\": rpc error: code = NotFound desc = could not find container \"1a898502c172251a58f921f830b7d6d6f8f096259c3dca6353d228972a83e528\": container with ID starting with 1a898502c172251a58f921f830b7d6d6f8f096259c3dca6353d228972a83e528 not found: ID does not exist" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.408672 1425 scope.go:115] "RemoveContainer" containerID="67e4592536523011a4901ffad309428f68583ceb75cd44ec7d261c76a0778640" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:14:13.408936 1425 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67e4592536523011a4901ffad309428f68583ceb75cd44ec7d261c76a0778640\": container with ID starting with 67e4592536523011a4901ffad309428f68583ceb75cd44ec7d261c76a0778640 not found: ID does not exist" containerID="67e4592536523011a4901ffad309428f68583ceb75cd44ec7d261c76a0778640" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.408970 1425 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:67e4592536523011a4901ffad309428f68583ceb75cd44ec7d261c76a0778640} err="failed to get container status \"67e4592536523011a4901ffad309428f68583ceb75cd44ec7d261c76a0778640\": rpc error: code = NotFound desc = could not find container \"67e4592536523011a4901ffad309428f68583ceb75cd44ec7d261c76a0778640\": container with ID starting with 67e4592536523011a4901ffad309428f68583ceb75cd44ec7d261c76a0778640 not found: ID does not exist" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.408983 1425 scope.go:115] "RemoveContainer" containerID="e6208ed7b9782cddc95b22490d2dc4f820f1944cb39050a5d478fabf458cd570" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:14:13.409200 1425 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6208ed7b9782cddc95b22490d2dc4f820f1944cb39050a5d478fabf458cd570\": container with ID starting with e6208ed7b9782cddc95b22490d2dc4f820f1944cb39050a5d478fabf458cd570 not found: ID does not exist" containerID="e6208ed7b9782cddc95b22490d2dc4f820f1944cb39050a5d478fabf458cd570" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.409238 1425 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:e6208ed7b9782cddc95b22490d2dc4f820f1944cb39050a5d478fabf458cd570} err="failed to get container status \"e6208ed7b9782cddc95b22490d2dc4f820f1944cb39050a5d478fabf458cd570\": rpc error: code = NotFound desc = could not find container \"e6208ed7b9782cddc95b22490d2dc4f820f1944cb39050a5d478fabf458cd570\": container with ID starting with e6208ed7b9782cddc95b22490d2dc4f820f1944cb39050a5d478fabf458cd570 not found: ID does not exist" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.409251 1425 scope.go:115] "RemoveContainer" containerID="976ee1fd7cc0d02f224dbcf7a41285bac05287d0f5cf4e79c3f6c098c2a7dcd4" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:14:13.409470 1425 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"976ee1fd7cc0d02f224dbcf7a41285bac05287d0f5cf4e79c3f6c098c2a7dcd4\": container with ID starting with 976ee1fd7cc0d02f224dbcf7a41285bac05287d0f5cf4e79c3f6c098c2a7dcd4 not found: ID does not exist" containerID="976ee1fd7cc0d02f224dbcf7a41285bac05287d0f5cf4e79c3f6c098c2a7dcd4" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.409499 1425 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:976ee1fd7cc0d02f224dbcf7a41285bac05287d0f5cf4e79c3f6c098c2a7dcd4} err="failed to get container status \"976ee1fd7cc0d02f224dbcf7a41285bac05287d0f5cf4e79c3f6c098c2a7dcd4\": rpc error: code = NotFound desc = could not find container \"976ee1fd7cc0d02f224dbcf7a41285bac05287d0f5cf4e79c3f6c098c2a7dcd4\": container with ID starting with 976ee1fd7cc0d02f224dbcf7a41285bac05287d0f5cf4e79c3f6c098c2a7dcd4 not found: ID does not exist" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.409514 1425 scope.go:115] "RemoveContainer" containerID="254f6adadc261bc8b8bef366241b799f8c167b2d7c165cb79c16c15ee634f029" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:14:13.409781 1425 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"254f6adadc261bc8b8bef366241b799f8c167b2d7c165cb79c16c15ee634f029\": container with ID starting with 254f6adadc261bc8b8bef366241b799f8c167b2d7c165cb79c16c15ee634f029 not found: ID does not exist" containerID="254f6adadc261bc8b8bef366241b799f8c167b2d7c165cb79c16c15ee634f029" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.409812 1425 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:254f6adadc261bc8b8bef366241b799f8c167b2d7c165cb79c16c15ee634f029} err="failed to get container status \"254f6adadc261bc8b8bef366241b799f8c167b2d7c165cb79c16c15ee634f029\": rpc error: code = NotFound desc = could not find container \"254f6adadc261bc8b8bef366241b799f8c167b2d7c165cb79c16c15ee634f029\": container with ID starting with 254f6adadc261bc8b8bef366241b799f8c167b2d7c165cb79c16c15ee634f029 not found: ID does not exist" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.409824 1425 scope.go:115] "RemoveContainer" containerID="6becde1d0c7f5d104d26d3a1c07efaa41bf091484c482dfffd15e87e4b5bf54c" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:14:13.410028 1425 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6becde1d0c7f5d104d26d3a1c07efaa41bf091484c482dfffd15e87e4b5bf54c\": container with ID starting with 6becde1d0c7f5d104d26d3a1c07efaa41bf091484c482dfffd15e87e4b5bf54c not found: ID does not exist" containerID="6becde1d0c7f5d104d26d3a1c07efaa41bf091484c482dfffd15e87e4b5bf54c" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.410059 1425 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:6becde1d0c7f5d104d26d3a1c07efaa41bf091484c482dfffd15e87e4b5bf54c} err="failed to get container status \"6becde1d0c7f5d104d26d3a1c07efaa41bf091484c482dfffd15e87e4b5bf54c\": rpc error: code = NotFound desc = could not find container \"6becde1d0c7f5d104d26d3a1c07efaa41bf091484c482dfffd15e87e4b5bf54c\": container with ID starting with 6becde1d0c7f5d104d26d3a1c07efaa41bf091484c482dfffd15e87e4b5bf54c not found: ID does not exist" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.410071 1425 scope.go:115] "RemoveContainer" containerID="c2fc2489b690ab18d10fc72750d26008cb07a65be8420998a29168a6a7cea2af" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:14:13.410309 1425 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2fc2489b690ab18d10fc72750d26008cb07a65be8420998a29168a6a7cea2af\": container with ID starting with c2fc2489b690ab18d10fc72750d26008cb07a65be8420998a29168a6a7cea2af not found: ID does not exist" containerID="c2fc2489b690ab18d10fc72750d26008cb07a65be8420998a29168a6a7cea2af" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.410337 1425 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:c2fc2489b690ab18d10fc72750d26008cb07a65be8420998a29168a6a7cea2af} err="failed to get container status \"c2fc2489b690ab18d10fc72750d26008cb07a65be8420998a29168a6a7cea2af\": rpc error: code = NotFound desc = could not find container \"c2fc2489b690ab18d10fc72750d26008cb07a65be8420998a29168a6a7cea2af\": container with ID starting with c2fc2489b690ab18d10fc72750d26008cb07a65be8420998a29168a6a7cea2af not found: ID does not exist" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.410350 1425 scope.go:115] "RemoveContainer" containerID="1a898502c172251a58f921f830b7d6d6f8f096259c3dca6353d228972a83e528" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.410628 1425 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:1a898502c172251a58f921f830b7d6d6f8f096259c3dca6353d228972a83e528} err="failed to get container status \"1a898502c172251a58f921f830b7d6d6f8f096259c3dca6353d228972a83e528\": rpc error: code = NotFound desc = could not find container \"1a898502c172251a58f921f830b7d6d6f8f096259c3dca6353d228972a83e528\": container with ID starting with 1a898502c172251a58f921f830b7d6d6f8f096259c3dca6353d228972a83e528 not found: ID does not exist" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.410645 1425 scope.go:115] "RemoveContainer" containerID="67e4592536523011a4901ffad309428f68583ceb75cd44ec7d261c76a0778640" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.410849 1425 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:67e4592536523011a4901ffad309428f68583ceb75cd44ec7d261c76a0778640} err="failed to get container status \"67e4592536523011a4901ffad309428f68583ceb75cd44ec7d261c76a0778640\": rpc error: code = NotFound desc = could not find container \"67e4592536523011a4901ffad309428f68583ceb75cd44ec7d261c76a0778640\": container with ID starting with 67e4592536523011a4901ffad309428f68583ceb75cd44ec7d261c76a0778640 not found: ID does not exist" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.410866 1425 scope.go:115] "RemoveContainer" containerID="e6208ed7b9782cddc95b22490d2dc4f820f1944cb39050a5d478fabf458cd570" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.411096 1425 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:e6208ed7b9782cddc95b22490d2dc4f820f1944cb39050a5d478fabf458cd570} err="failed to get container status \"e6208ed7b9782cddc95b22490d2dc4f820f1944cb39050a5d478fabf458cd570\": rpc error: code = NotFound desc = could not find container \"e6208ed7b9782cddc95b22490d2dc4f820f1944cb39050a5d478fabf458cd570\": container with ID starting with e6208ed7b9782cddc95b22490d2dc4f820f1944cb39050a5d478fabf458cd570 not found: ID does not exist" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.411110 1425 scope.go:115] "RemoveContainer" containerID="976ee1fd7cc0d02f224dbcf7a41285bac05287d0f5cf4e79c3f6c098c2a7dcd4" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.411334 1425 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:976ee1fd7cc0d02f224dbcf7a41285bac05287d0f5cf4e79c3f6c098c2a7dcd4} err="failed to get container status \"976ee1fd7cc0d02f224dbcf7a41285bac05287d0f5cf4e79c3f6c098c2a7dcd4\": rpc error: code = NotFound desc = could not find container \"976ee1fd7cc0d02f224dbcf7a41285bac05287d0f5cf4e79c3f6c098c2a7dcd4\": container with ID starting with 976ee1fd7cc0d02f224dbcf7a41285bac05287d0f5cf4e79c3f6c098c2a7dcd4 not found: ID does not exist" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.411350 1425 scope.go:115] "RemoveContainer" containerID="254f6adadc261bc8b8bef366241b799f8c167b2d7c165cb79c16c15ee634f029" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.411629 1425 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:254f6adadc261bc8b8bef366241b799f8c167b2d7c165cb79c16c15ee634f029} err="failed to get container status \"254f6adadc261bc8b8bef366241b799f8c167b2d7c165cb79c16c15ee634f029\": rpc error: code = NotFound desc = could not find container \"254f6adadc261bc8b8bef366241b799f8c167b2d7c165cb79c16c15ee634f029\": container with ID starting with 254f6adadc261bc8b8bef366241b799f8c167b2d7c165cb79c16c15ee634f029 not found: ID does not exist" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.411643 1425 scope.go:115] "RemoveContainer" containerID="6becde1d0c7f5d104d26d3a1c07efaa41bf091484c482dfffd15e87e4b5bf54c" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.411857 1425 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:6becde1d0c7f5d104d26d3a1c07efaa41bf091484c482dfffd15e87e4b5bf54c} err="failed to get container status \"6becde1d0c7f5d104d26d3a1c07efaa41bf091484c482dfffd15e87e4b5bf54c\": rpc error: code = NotFound desc = could not find container \"6becde1d0c7f5d104d26d3a1c07efaa41bf091484c482dfffd15e87e4b5bf54c\": container with ID starting with 6becde1d0c7f5d104d26d3a1c07efaa41bf091484c482dfffd15e87e4b5bf54c not found: ID does not exist" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.411876 1425 scope.go:115] "RemoveContainer" containerID="c2fc2489b690ab18d10fc72750d26008cb07a65be8420998a29168a6a7cea2af" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.412106 1425 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:c2fc2489b690ab18d10fc72750d26008cb07a65be8420998a29168a6a7cea2af} err="failed to get container status \"c2fc2489b690ab18d10fc72750d26008cb07a65be8420998a29168a6a7cea2af\": rpc error: code = NotFound desc = could not find container \"c2fc2489b690ab18d10fc72750d26008cb07a65be8420998a29168a6a7cea2af\": container with ID starting with c2fc2489b690ab18d10fc72750d26008cb07a65be8420998a29168a6a7cea2af not found: ID does not exist" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.412127 1425 scope.go:115] "RemoveContainer" containerID="a30b14e5cf7f780de074100612e1db98e849c8b763e7b551ab0f6558754f69f6" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:13.420503467Z" level=info msg="Removing container: a30b14e5cf7f780de074100612e1db98e849c8b763e7b551ab0f6558754f69f6" id=87c50804-e8cc-40d2-b106-b565f7765cfd name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:13.439325038Z" level=info msg="Removed container a30b14e5cf7f780de074100612e1db98e849c8b763e7b551ab0f6558754f69f6: openshift-monitoring/prometheus-adapter-5894cc86c7-47422/prometheus-adapter" id=87c50804-e8cc-40d2-b106-b565f7765cfd name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.439518 1425 scope.go:115] "RemoveContainer" containerID="a30b14e5cf7f780de074100612e1db98e849c8b763e7b551ab0f6558754f69f6" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:14:13.439814 1425 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a30b14e5cf7f780de074100612e1db98e849c8b763e7b551ab0f6558754f69f6\": container with ID starting with a30b14e5cf7f780de074100612e1db98e849c8b763e7b551ab0f6558754f69f6 not found: ID does not exist" containerID="a30b14e5cf7f780de074100612e1db98e849c8b763e7b551ab0f6558754f69f6" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.439849 1425 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:a30b14e5cf7f780de074100612e1db98e849c8b763e7b551ab0f6558754f69f6} err="failed to get container status \"a30b14e5cf7f780de074100612e1db98e849c8b763e7b551ab0f6558754f69f6\": rpc error: code = NotFound desc = could not find container \"a30b14e5cf7f780de074100612e1db98e849c8b763e7b551ab0f6558754f69f6\": container with ID starting with a30b14e5cf7f780de074100612e1db98e849c8b763e7b551ab0f6558754f69f6 not found: ID does not exist" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.439863 1425 scope.go:115] "RemoveContainer" containerID="b6613cda35aa605c0c2f9fa03950bbd6def356ba0a1a95a665f8c80d6b17a14f" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:13.440538255Z" level=info msg="Removing container: b6613cda35aa605c0c2f9fa03950bbd6def356ba0a1a95a665f8c80d6b17a14f" id=2fd317f1-a954-45ec-aca0-daac24de247a name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:13.456874863Z" level=info msg="Removed container b6613cda35aa605c0c2f9fa03950bbd6def356ba0a1a95a665f8c80d6b17a14f: openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-54rjq/prometheus-operator-admission-webhook" id=2fd317f1-a954-45ec-aca0-daac24de247a name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.457042 1425 scope.go:115] "RemoveContainer" containerID="3b8c93b35afc863ee93d43a145f66340ca0ab0a562aca1259bff690246d320cf" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:13.457730637Z" level=info msg="Removing container: 3b8c93b35afc863ee93d43a145f66340ca0ab0a562aca1259bff690246d320cf" id=106c4b05-36e0-4c6f-92a8-17e4a0da26cb name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.482203 1425 reconciler.go:399] "Volume detached for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-secret-alertmanager-kube-rbac-proxy-metric\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.482238 1425 reconciler.go:399] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-web-config\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.482260 1425 reconciler.go:399] "Volume detached for volume \"secret-alertmanager-main-proxy\" (UniqueName: \"kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-secret-alertmanager-main-proxy\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.482278 1425 reconciler.go:399] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-config-volume\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.482300 1425 reconciler.go:399] "Volume detached for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-secret-alertmanager-main-tls\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.482318 1425 reconciler.go:399] "Volume detached for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/b59f4808-02df-4b2e-9694-8cf72a1b5e75-metrics-client-ca\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.482337 1425 reconciler.go:399] "Volume detached for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/b59f4808-02df-4b2e-9694-8cf72a1b5e75-alertmanager-main-db\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.482354 1425 reconciler.go:399] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b59f4808-02df-4b2e-9694-8cf72a1b5e75-config-out\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.482372 1425 reconciler.go:399] "Volume detached for volume \"kube-api-access-xqgp9\" (UniqueName: \"kubernetes.io/projected/b59f4808-02df-4b2e-9694-8cf72a1b5e75-kube-api-access-xqgp9\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.482413 1425 reconciler.go:399] "Volume detached for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/b59f4808-02df-4b2e-9694-8cf72a1b5e75-secret-alertmanager-kube-rbac-proxy\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.482432 1425 reconciler.go:399] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b59f4808-02df-4b2e-9694-8cf72a1b5e75-tls-assets\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.482451 1425 reconciler.go:399] "Volume detached for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b59f4808-02df-4b2e-9694-8cf72a1b5e75-alertmanager-trusted-ca-bundle\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:13.485612025Z" level=info msg="Removed container 3b8c93b35afc863ee93d43a145f66340ca0ab0a562aca1259bff690246d320cf: openshift-network-diagnostics/network-check-source-6775957967-zsr9c/check-endpoints" id=106c4b05-36e0-4c6f-92a8-17e4a0da26cb name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.781528 1425 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-6f4f5fb7cf-wtcwx_72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c/oauth-proxy/0.log" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.782095 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx" event=&{ID:72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c Type:ContainerDied Data:6c2cdef5be97661e24b16499c5f84baa5f0fdffbceb45f19b515c50e93bb6dd5} Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.782132 1425 scope.go:115] "RemoveContainer" containerID="2bce691bc41bddf8da3b590cbdbb6b3b7604a9e34c167c7d2a14b15a944439f1" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:13.784921460Z" level=info msg="Removing container: 2bce691bc41bddf8da3b590cbdbb6b3b7604a9e34c167c7d2a14b15a944439f1" id=6b404549-e5dd-4275-ba32-4e5db0798b08 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.805526 1425 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-1_b59f4808-02df-4b2e-9694-8cf72a1b5e75/alertmanager-proxy/0.log" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.806596 1425 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-1_b59f4808-02df-4b2e-9694-8cf72a1b5e75/config-reloader/0.log" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.808434 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-1" event=&{ID:b59f4808-02df-4b2e-9694-8cf72a1b5e75 Type:ContainerDied Data:4cc6b8c0009aed4f5500066da7f716df203dd1969281bb8dbce4d7052e55ec50} Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:13.853906242Z" level=info msg="Removed container 2bce691bc41bddf8da3b590cbdbb6b3b7604a9e34c167c7d2a14b15a944439f1: openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/kube-rbac-proxy-metrics" id=6b404549-e5dd-4275-ba32-4e5db0798b08 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.854223 1425 scope.go:115] "RemoveContainer" containerID="aa7b352bf36356466f81f85608b01622b4f84956d931d3153417ad8abf7d95e9" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:13.855263236Z" level=info msg="Removing container: aa7b352bf36356466f81f85608b01622b4f84956d931d3153417ad8abf7d95e9" id=44721239-0704-453e-b905-a7f1251d18ca name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.864933 1425 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-monitoring/alertmanager-main-1] Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.873359 1425 kubelet.go:2129] "SyncLoop REMOVE" source="api" pods=[openshift-monitoring/alertmanager-main-1] Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:13.890693165Z" level=info msg="Removed container aa7b352bf36356466f81f85608b01622b4f84956d931d3153417ad8abf7d95e9: openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/kube-rbac-proxy-rules" id=44721239-0704-453e-b905-a7f1251d18ca name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:13.891706905Z" level=info msg="Removing container: 81147f8c4de8241e926f4c84fcb3c6b8d438e973b16a73d5c29483986c87857b" id=b669acbd-c452-4eb7-b532-bd4df79dd68a name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.890879 1425 scope.go:115] "RemoveContainer" containerID="81147f8c4de8241e926f4c84fcb3c6b8d438e973b16a73d5c29483986c87857b" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.895532 1425 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx] Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.909148 1425 kubelet.go:2129] "SyncLoop REMOVE" source="api" pods=[openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx] Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:13.937827734Z" level=info msg="Removed container 81147f8c4de8241e926f4c84fcb3c6b8d438e973b16a73d5c29483986c87857b: openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/prom-label-proxy" id=b669acbd-c452-4eb7-b532-bd4df79dd68a name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.938886 1425 scope.go:115] "RemoveContainer" containerID="217b93493832653788189b59ae04c4259e18df93a8c0c89eb6d7197fdb6bdf27" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:13.939816146Z" level=info msg="Removing container: 217b93493832653788189b59ae04c4259e18df93a8c0c89eb6d7197fdb6bdf27" id=7fd441ac-a22f-43d1-a524-52a9742ceb66 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:13.967692644Z" level=info msg="Removed container 217b93493832653788189b59ae04c4259e18df93a8c0c89eb6d7197fdb6bdf27: openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/kube-rbac-proxy" id=7fd441ac-a22f-43d1-a524-52a9742ceb66 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:13.968707 1425 scope.go:115] "RemoveContainer" containerID="e67ece935b56afbeb827b9311beb6a1f9f4afc8db2691bdcd603bd752c6deaa7" Jan 05 09:14:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:13.970082968Z" level=info msg="Removing container: e67ece935b56afbeb827b9311beb6a1f9f4afc8db2691bdcd603bd752c6deaa7" id=375af4df-0835-4c2b-a43f-22a32815292e name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:14.014533091Z" level=info msg="Removed container e67ece935b56afbeb827b9311beb6a1f9f4afc8db2691bdcd603bd752c6deaa7: openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/oauth-proxy" id=375af4df-0835-4c2b-a43f-22a32815292e name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:14.014856 1425 scope.go:115] "RemoveContainer" containerID="fcb76a52605ae84e7415c0b6e56aa8e1f010607edc042a479cb6e4adde24ecc4" Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:14.020444689Z" level=info msg="Removing container: fcb76a52605ae84e7415c0b6e56aa8e1f010607edc042a479cb6e4adde24ecc4" id=7f38380a-71ba-4b78-ab9c-deb774efd208 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:14.056875368Z" level=info msg="Removed container fcb76a52605ae84e7415c0b6e56aa8e1f010607edc042a479cb6e4adde24ecc4: openshift-monitoring/thanos-querier-6f4f5fb7cf-wtcwx/thanos-query" id=7f38380a-71ba-4b78-ab9c-deb774efd208 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:14.057219 1425 scope.go:115] "RemoveContainer" containerID="b9bc795189fd458b4ce14c6f192418f946a1c35d2f19ab6aedd4b4185f4bf487" Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:14.058408429Z" level=info msg="Removing container: b9bc795189fd458b4ce14c6f192418f946a1c35d2f19ab6aedd4b4185f4bf487" id=66b99378-39e0-4a46-bc0e-3c25a7a32016 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:14.091987934Z" level=info msg="Removed container b9bc795189fd458b4ce14c6f192418f946a1c35d2f19ab6aedd4b4185f4bf487: openshift-monitoring/alertmanager-main-1/prom-label-proxy" id=66b99378-39e0-4a46-bc0e-3c25a7a32016 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:14.093067 1425 scope.go:115] "RemoveContainer" containerID="064809cddb0902a863e4fcd8d1da29b6a436aa3118d40b3b5ab6fbbda4fbf3d9" Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:14.094408156Z" level=info msg="Removing container: 064809cddb0902a863e4fcd8d1da29b6a436aa3118d40b3b5ab6fbbda4fbf3d9" id=0959d2da-4623-4919-a671-c721bc67e10f name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:14.123960067Z" level=info msg="Removed container 064809cddb0902a863e4fcd8d1da29b6a436aa3118d40b3b5ab6fbbda4fbf3d9: openshift-monitoring/alertmanager-main-1/kube-rbac-proxy-metric" id=0959d2da-4623-4919-a671-c721bc67e10f name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:14.124460 1425 scope.go:115] "RemoveContainer" containerID="6eba543bd270671b29706b350f4dfdf034230e4d33c811838d47249838c4ff2b" Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:14.125503150Z" level=info msg="Removing container: 6eba543bd270671b29706b350f4dfdf034230e4d33c811838d47249838c4ff2b" id=463e7e9f-99bf-408c-9e24-6399e3ff6c0c name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:14.183028874Z" level=info msg="Removed container 6eba543bd270671b29706b350f4dfdf034230e4d33c811838d47249838c4ff2b: openshift-monitoring/alertmanager-main-1/kube-rbac-proxy" id=463e7e9f-99bf-408c-9e24-6399e3ff6c0c name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:14.183569 1425 scope.go:115] "RemoveContainer" containerID="af5d20453c1aa7ea8d99d15e82767c98893d319f38bb9cd97aa815202e28d32c" Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:14.184596151Z" level=info msg="Removing container: af5d20453c1aa7ea8d99d15e82767c98893d319f38bb9cd97aa815202e28d32c" id=34ebfc72-0499-41f1-b33f-1108380ba299 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:14.220747579Z" level=info msg="Removed container af5d20453c1aa7ea8d99d15e82767c98893d319f38bb9cd97aa815202e28d32c: openshift-monitoring/alertmanager-main-1/alertmanager-proxy" id=34ebfc72-0499-41f1-b33f-1108380ba299 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:14.221035 1425 scope.go:115] "RemoveContainer" containerID="e9efd1bb280b0736df20aa595c1c6f65c0d953e8a7a6a1dde105be0f728d6de6" Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:14.221989817Z" level=info msg="Removing container: e9efd1bb280b0736df20aa595c1c6f65c0d953e8a7a6a1dde105be0f728d6de6" id=8198f28a-4930-4bea-bb5e-61934879d0b0 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:14.254506592Z" level=info msg="Removed container e9efd1bb280b0736df20aa595c1c6f65c0d953e8a7a6a1dde105be0f728d6de6: openshift-monitoring/alertmanager-main-1/config-reloader" id=8198f28a-4930-4bea-bb5e-61934879d0b0 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:14.254739 1425 scope.go:115] "RemoveContainer" containerID="902fedd514d362305b82903fec325b5adb139bd729665753e388c28b1e4d58bf" Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:14.255447807Z" level=info msg="Removing container: 902fedd514d362305b82903fec325b5adb139bd729665753e388c28b1e4d58bf" id=26d57eee-ef48-4600-a3df-ec56c965413b name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:14.270374408Z" level=info msg="Removed container 902fedd514d362305b82903fec325b5adb139bd729665753e388c28b1e4d58bf: openshift-monitoring/alertmanager-main-1/alertmanager" id=26d57eee-ef48-4600-a3df-ec56c965413b name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:14.664489776Z" level=info msg="Stopping pod sandbox: 4cc6b8c0009aed4f5500066da7f716df203dd1969281bb8dbce4d7052e55ec50" id=f957ca48-2b17-4d6e-ba1c-e9a7810422ad name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:14.664533155Z" level=info msg="Stopped pod sandbox (already stopped): 4cc6b8c0009aed4f5500066da7f716df203dd1969281bb8dbce4d7052e55ec50" id=f957ca48-2b17-4d6e-ba1c-e9a7810422ad name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:14.664608913Z" level=info msg="Stopping pod sandbox: 6c2cdef5be97661e24b16499c5f84baa5f0fdffbceb45f19b515c50e93bb6dd5" id=98389c6d-5a29-4b6f-83a4-e8b9e5c666ad name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:14.664634696Z" level=info msg="Stopped pod sandbox (already stopped): 6c2cdef5be97661e24b16499c5f84baa5f0fdffbceb45f19b515c50e93bb6dd5" id=98389c6d-5a29-4b6f-83a4-e8b9e5c666ad name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:14.666432 1425 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c path="/var/lib/kubelet/pods/72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c/volumes" Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:14.667352 1425 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=89a48689-810c-464e-9ac6-d78977e60793 path="/var/lib/kubelet/pods/89a48689-810c-464e-9ac6-d78977e60793/volumes" Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:14.668019 1425 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=8ac551e8-a959-4212-86a1-2b3ac482586f path="/var/lib/kubelet/pods/8ac551e8-a959-4212-86a1-2b3ac482586f/volumes" Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:14.669115 1425 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=a82fd83b-5344-4e7a-91dd-5894f3846e87 path="/var/lib/kubelet/pods/a82fd83b-5344-4e7a-91dd-5894f3846e87/volumes" Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:14.669794 1425 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=b59f4808-02df-4b2e-9694-8cf72a1b5e75 path="/var/lib/kubelet/pods/b59f4808-02df-4b2e-9694-8cf72a1b5e75/volumes" Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:14.670693 1425 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=f8e5be3f-9ab7-484c-b61b-6231a8774e24 path="/var/lib/kubelet/pods/f8e5be3f-9ab7-484c-b61b-6231a8774e24/volumes" Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:14.760564 1425 patch_prober.go:29] interesting pod/router-default-5cf6bbdff9-fkmcn container/router namespace/openshift-ingress: Readiness probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]backend-http ok Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: [+]has-synced ok Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: [-]process-running failed: reason withheld Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: healthz check failed Jan 05 09:14:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:14.760617 1425 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" podUID=66d22b70-08bb-4199-b258-ac4590e9b487 containerName="router" probeResult=failure output="HTTP probe failed with statuscode: 500" Jan 05 09:14:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:18.420476 1425 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeNotSchedulable" Jan 05 09:14:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:24.759910 1425 patch_prober.go:29] interesting pod/router-default-5cf6bbdff9-fkmcn container/router namespace/openshift-ingress: Readiness probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]backend-http ok Jan 05 09:14:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: [+]has-synced ok Jan 05 09:14:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: [-]process-running failed: reason withheld Jan 05 09:14:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: healthz check failed Jan 05 09:14:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:24.759967 1425 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" podUID=66d22b70-08bb-4199-b258-ac4590e9b487 containerName="router" probeResult=failure output="HTTP probe failed with statuscode: 500" Jan 05 09:14:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:34.760056 1425 patch_prober.go:29] interesting pod/router-default-5cf6bbdff9-fkmcn container/router namespace/openshift-ingress: Readiness probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]backend-http ok Jan 05 09:14:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: [+]has-synced ok Jan 05 09:14:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: [-]process-running failed: reason withheld Jan 05 09:14:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: healthz check failed Jan 05 09:14:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:34.760128 1425 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" podUID=66d22b70-08bb-4199-b258-ac4590e9b487 containerName="router" probeResult=failure output="HTTP probe failed with statuscode: 500" Jan 05 09:14:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:34.760211 1425 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" Jan 05 09:14:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:36.057894 1425 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-image-registry/image-registry-64655f6865-dx2q6" podUID=4685a26b-f07f-4b10-8547-dea3d6802723 containerName="registry" containerID="cri-o://784e9766ec3aebd02560a716704c4d48760309927d4008e2c136d460249ddb98" gracePeriod=55 Jan 05 09:14:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:36.058449747Z" level=info msg="Stopping container: 784e9766ec3aebd02560a716704c4d48760309927d4008e2c136d460249ddb98 (timeout: 55s)" id=73406110-73d1-4681-82f1-45a60bdeaf40 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:36.673918344Z" level=info msg="Stopping pod sandbox: 148c6bc4da65fd9b07c5fca9d24c7096ee2b924f8d407958ac234e9758e340f4" id=bd0cd92b-9c56-4454-8411-dc42965e8535 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:36.673959048Z" level=info msg="Stopped pod sandbox (already stopped): 148c6bc4da65fd9b07c5fca9d24c7096ee2b924f8d407958ac234e9758e340f4" id=bd0cd92b-9c56-4454-8411-dc42965e8535 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:36.674308166Z" level=info msg="Removing pod sandbox: 148c6bc4da65fd9b07c5fca9d24c7096ee2b924f8d407958ac234e9758e340f4" id=36e2eaa2-6cf2-414c-9569-d4a410eb2cfa name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:14:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:36.681790230Z" level=info msg="Removed pod sandbox: 148c6bc4da65fd9b07c5fca9d24c7096ee2b924f8d407958ac234e9758e340f4" id=36e2eaa2-6cf2-414c-9569-d4a410eb2cfa name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:14:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:36.682030123Z" level=info msg="Stopping pod sandbox: 4cc6b8c0009aed4f5500066da7f716df203dd1969281bb8dbce4d7052e55ec50" id=a9a38301-6c23-4690-a602-9b35c42a2eb8 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:36.682061530Z" level=info msg="Stopped pod sandbox (already stopped): 4cc6b8c0009aed4f5500066da7f716df203dd1969281bb8dbce4d7052e55ec50" id=a9a38301-6c23-4690-a602-9b35c42a2eb8 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:36.682317477Z" level=info msg="Removing pod sandbox: 4cc6b8c0009aed4f5500066da7f716df203dd1969281bb8dbce4d7052e55ec50" id=71083be8-aa3e-4d73-ad05-d29bfcf9a108 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:14:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:36.689082230Z" level=info msg="Removed pod sandbox: 4cc6b8c0009aed4f5500066da7f716df203dd1969281bb8dbce4d7052e55ec50" id=71083be8-aa3e-4d73-ad05-d29bfcf9a108 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:14:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:36.689299886Z" level=info msg="Stopping pod sandbox: e7532666db14128bc848af4cbb52adc603db1ec585b807063b42799df403a10b" id=3e502324-7578-4fd0-80b5-1004fc20242b name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:36.689329861Z" level=info msg="Stopped pod sandbox (already stopped): e7532666db14128bc848af4cbb52adc603db1ec585b807063b42799df403a10b" id=3e502324-7578-4fd0-80b5-1004fc20242b name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:36.689625302Z" level=info msg="Removing pod sandbox: e7532666db14128bc848af4cbb52adc603db1ec585b807063b42799df403a10b" id=33b5f397-54d9-4212-a4a1-d45898db0419 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:14:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:36.695452953Z" level=info msg="Removed pod sandbox: e7532666db14128bc848af4cbb52adc603db1ec585b807063b42799df403a10b" id=33b5f397-54d9-4212-a4a1-d45898db0419 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:14:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:36.695688376Z" level=info msg="Stopping pod sandbox: 8e21977f4f2aa77ecd7154452d4a15adc91d3b323c169dda297ad9c1f8901298" id=a7a5a71e-4c61-419b-a41e-b6fd97b3bde7 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:36.695721726Z" level=info msg="Stopped pod sandbox (already stopped): 8e21977f4f2aa77ecd7154452d4a15adc91d3b323c169dda297ad9c1f8901298" id=a7a5a71e-4c61-419b-a41e-b6fd97b3bde7 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:36.695929810Z" level=info msg="Removing pod sandbox: 8e21977f4f2aa77ecd7154452d4a15adc91d3b323c169dda297ad9c1f8901298" id=47a392b8-282f-461a-958d-4518096db378 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:14:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:36.709102322Z" level=info msg="Removed pod sandbox: 8e21977f4f2aa77ecd7154452d4a15adc91d3b323c169dda297ad9c1f8901298" id=47a392b8-282f-461a-958d-4518096db378 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:14:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:36.709328439Z" level=info msg="Stopping pod sandbox: 6c2cdef5be97661e24b16499c5f84baa5f0fdffbceb45f19b515c50e93bb6dd5" id=c59f8b83-e2d4-4621-8dc3-1dfe67d8f9fe name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:36.709361028Z" level=info msg="Stopped pod sandbox (already stopped): 6c2cdef5be97661e24b16499c5f84baa5f0fdffbceb45f19b515c50e93bb6dd5" id=c59f8b83-e2d4-4621-8dc3-1dfe67d8f9fe name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:36.709636386Z" level=info msg="Removing pod sandbox: 6c2cdef5be97661e24b16499c5f84baa5f0fdffbceb45f19b515c50e93bb6dd5" id=234abe72-072a-461b-9d93-2b1dd19e3ede name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:14:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:36.715441776Z" level=info msg="Removed pod sandbox: 6c2cdef5be97661e24b16499c5f84baa5f0fdffbceb45f19b515c50e93bb6dd5" id=234abe72-072a-461b-9d93-2b1dd19e3ede name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:14:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:36.715705067Z" level=info msg="Stopping pod sandbox: bf998995ead8afb2642821b82636d7b867d8c0aa5ab164fe09e7396c2701f63b" id=3680cbc1-7b35-419a-b006-3d57322e9e4d name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:36.715727586Z" level=info msg="Stopped pod sandbox (already stopped): bf998995ead8afb2642821b82636d7b867d8c0aa5ab164fe09e7396c2701f63b" id=3680cbc1-7b35-419a-b006-3d57322e9e4d name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:36.715934199Z" level=info msg="Removing pod sandbox: bf998995ead8afb2642821b82636d7b867d8c0aa5ab164fe09e7396c2701f63b" id=557c309f-50c1-4d35-8eb8-71940a78a7b1 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:14:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:36.721718596Z" level=info msg="Removed pod sandbox: bf998995ead8afb2642821b82636d7b867d8c0aa5ab164fe09e7396c2701f63b" id=557c309f-50c1-4d35-8eb8-71940a78a7b1 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:14:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:36.721961972Z" level=info msg="Stopping pod sandbox: ef7044b787198dfb2aab7533e9c72c07b8c2ef842ecb652a788dc647c485157c" id=1fcf912b-f81a-46a8-9361-f9ebae7fa993 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:36.721991157Z" level=info msg="Stopped pod sandbox (already stopped): ef7044b787198dfb2aab7533e9c72c07b8c2ef842ecb652a788dc647c485157c" id=1fcf912b-f81a-46a8-9361-f9ebae7fa993 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:36.722217002Z" level=info msg="Removing pod sandbox: ef7044b787198dfb2aab7533e9c72c07b8c2ef842ecb652a788dc647c485157c" id=a70b793a-7f8e-4cf8-ba41-61de66804b6d name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:14:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:36.728218720Z" level=info msg="Removed pod sandbox: ef7044b787198dfb2aab7533e9c72c07b8c2ef842ecb652a788dc647c485157c" id=a70b793a-7f8e-4cf8-ba41-61de66804b6d name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:37.206739726Z" level=info msg="Stopped container 784e9766ec3aebd02560a716704c4d48760309927d4008e2c136d460249ddb98: openshift-image-registry/image-registry-64655f6865-dx2q6/registry" id=73406110-73d1-4681-82f1-45a60bdeaf40 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:37.207350097Z" level=info msg="Stopping pod sandbox: f855fd38c50954c79bc8f0f3e034b1d0f7293735e9dbc194b9e271cfb674c01c" id=7c3cb772-58bd-478a-a780-4b99050394bf name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:37.207585008Z" level=info msg="Got pod network &{Name:image-registry-64655f6865-dx2q6 Namespace:openshift-image-registry ID:f855fd38c50954c79bc8f0f3e034b1d0f7293735e9dbc194b9e271cfb674c01c UID:4685a26b-f07f-4b10-8547-dea3d6802723 NetNS:/var/run/netns/c8c45494-0105-46cf-aadf-e33e0bf7b218 Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:37.207695329Z" level=info msg="Deleting pod openshift-image-registry_image-registry-64655f6865-dx2q6 from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: 2023-01-05T09:14:37Z [verbose] Del: openshift-image-registry:image-registry-64655f6865-dx2q6:4685a26b-f07f-4b10-8547-dea3d6802723:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:37.429801683Z" level=info msg="Stopped pod sandbox: f855fd38c50954c79bc8f0f3e034b1d0f7293735e9dbc194b9e271cfb674c01c" id=7c3cb772-58bd-478a-a780-4b99050394bf name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.547810 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4685a26b-f07f-4b10-8547-dea3d6802723-registry-tls\") pod \"4685a26b-f07f-4b10-8547-dea3d6802723\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.547859 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"kube-api-access-db7jj\" (UniqueName: \"kubernetes.io/projected/4685a26b-f07f-4b10-8547-dea3d6802723-kube-api-access-db7jj\") pod \"4685a26b-f07f-4b10-8547-dea3d6802723\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.547894 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4685a26b-f07f-4b10-8547-dea3d6802723-ca-trust-extracted\") pod \"4685a26b-f07f-4b10-8547-dea3d6802723\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.547922 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4685a26b-f07f-4b10-8547-dea3d6802723-trusted-ca\") pod \"4685a26b-f07f-4b10-8547-dea3d6802723\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.547959 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4685a26b-f07f-4b10-8547-dea3d6802723-installation-pull-secrets\") pod \"4685a26b-f07f-4b10-8547-dea3d6802723\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.547990 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4685a26b-f07f-4b10-8547-dea3d6802723-registry-certificates\") pod \"4685a26b-f07f-4b10-8547-dea3d6802723\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.548022 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4685a26b-f07f-4b10-8547-dea3d6802723-bound-sa-token\") pod \"4685a26b-f07f-4b10-8547-dea3d6802723\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.548069 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/4685a26b-f07f-4b10-8547-dea3d6802723-image-registry-private-configuration\") pod \"4685a26b-f07f-4b10-8547-dea3d6802723\" (UID: \"4685a26b-f07f-4b10-8547-dea3d6802723\") " Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:14:37.548717 1425 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/4685a26b-f07f-4b10-8547-dea3d6802723/volumes/kubernetes.io~configmap/trusted-ca: clearQuota called, but quotas disabled Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.549005 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4685a26b-f07f-4b10-8547-dea3d6802723-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "4685a26b-f07f-4b10-8547-dea3d6802723" (UID: "4685a26b-f07f-4b10-8547-dea3d6802723"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:14:37.549532 1425 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/4685a26b-f07f-4b10-8547-dea3d6802723/volumes/kubernetes.io~empty-dir/ca-trust-extracted: clearQuota called, but quotas disabled Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.550156 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4685a26b-f07f-4b10-8547-dea3d6802723-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "4685a26b-f07f-4b10-8547-dea3d6802723" (UID: "4685a26b-f07f-4b10-8547-dea3d6802723"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:14:37.550306 1425 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/4685a26b-f07f-4b10-8547-dea3d6802723/volumes/kubernetes.io~configmap/registry-certificates: clearQuota called, but quotas disabled Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.550509 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4685a26b-f07f-4b10-8547-dea3d6802723-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "4685a26b-f07f-4b10-8547-dea3d6802723" (UID: "4685a26b-f07f-4b10-8547-dea3d6802723"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.553968 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4685a26b-f07f-4b10-8547-dea3d6802723-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "4685a26b-f07f-4b10-8547-dea3d6802723" (UID: "4685a26b-f07f-4b10-8547-dea3d6802723"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.559643 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4685a26b-f07f-4b10-8547-dea3d6802723-image-registry-private-configuration" (OuterVolumeSpecName: "image-registry-private-configuration") pod "4685a26b-f07f-4b10-8547-dea3d6802723" (UID: "4685a26b-f07f-4b10-8547-dea3d6802723"). InnerVolumeSpecName "image-registry-private-configuration". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.572021 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4685a26b-f07f-4b10-8547-dea3d6802723-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "4685a26b-f07f-4b10-8547-dea3d6802723" (UID: "4685a26b-f07f-4b10-8547-dea3d6802723"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.572032 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4685a26b-f07f-4b10-8547-dea3d6802723-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "4685a26b-f07f-4b10-8547-dea3d6802723" (UID: "4685a26b-f07f-4b10-8547-dea3d6802723"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.572381 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4685a26b-f07f-4b10-8547-dea3d6802723-kube-api-access-db7jj" (OuterVolumeSpecName: "kube-api-access-db7jj") pod "4685a26b-f07f-4b10-8547-dea3d6802723" (UID: "4685a26b-f07f-4b10-8547-dea3d6802723"). InnerVolumeSpecName "kube-api-access-db7jj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.648748 1425 reconciler.go:399] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4685a26b-f07f-4b10-8547-dea3d6802723-registry-tls\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.648781 1425 reconciler.go:399] "Volume detached for volume \"kube-api-access-db7jj\" (UniqueName: \"kubernetes.io/projected/4685a26b-f07f-4b10-8547-dea3d6802723-kube-api-access-db7jj\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.648797 1425 reconciler.go:399] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4685a26b-f07f-4b10-8547-dea3d6802723-ca-trust-extracted\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.648812 1425 reconciler.go:399] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4685a26b-f07f-4b10-8547-dea3d6802723-trusted-ca\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.648830 1425 reconciler.go:399] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4685a26b-f07f-4b10-8547-dea3d6802723-installation-pull-secrets\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.648844 1425 reconciler.go:399] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4685a26b-f07f-4b10-8547-dea3d6802723-registry-certificates\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.648858 1425 reconciler.go:399] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4685a26b-f07f-4b10-8547-dea3d6802723-bound-sa-token\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.648874 1425 reconciler.go:399] "Volume detached for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/4685a26b-f07f-4b10-8547-dea3d6802723-image-registry-private-configuration\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.889591 1425 generic.go:296] "Generic (PLEG): container finished" podID=4685a26b-f07f-4b10-8547-dea3d6802723 containerID="784e9766ec3aebd02560a716704c4d48760309927d4008e2c136d460249ddb98" exitCode=0 Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.889627 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-64655f6865-dx2q6" event=&{ID:4685a26b-f07f-4b10-8547-dea3d6802723 Type:ContainerDied Data:784e9766ec3aebd02560a716704c4d48760309927d4008e2c136d460249ddb98} Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.889650 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-64655f6865-dx2q6" event=&{ID:4685a26b-f07f-4b10-8547-dea3d6802723 Type:ContainerDied Data:f855fd38c50954c79bc8f0f3e034b1d0f7293735e9dbc194b9e271cfb674c01c} Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.889669 1425 scope.go:115] "RemoveContainer" containerID="784e9766ec3aebd02560a716704c4d48760309927d4008e2c136d460249ddb98" Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:37.894962710Z" level=info msg="Removing container: 784e9766ec3aebd02560a716704c4d48760309927d4008e2c136d460249ddb98" id=cedbe06d-243d-42d3-adb9-2d3b4caaeecf name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.919059 1425 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-image-registry/image-registry-64655f6865-dx2q6] Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:37.920470384Z" level=info msg="Removed container 784e9766ec3aebd02560a716704c4d48760309927d4008e2c136d460249ddb98: openshift-image-registry/image-registry-64655f6865-dx2q6/registry" id=cedbe06d-243d-42d3-adb9-2d3b4caaeecf name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.920698 1425 scope.go:115] "RemoveContainer" containerID="784e9766ec3aebd02560a716704c4d48760309927d4008e2c136d460249ddb98" Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:14:37.921011 1425 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"784e9766ec3aebd02560a716704c4d48760309927d4008e2c136d460249ddb98\": container with ID starting with 784e9766ec3aebd02560a716704c4d48760309927d4008e2c136d460249ddb98 not found: ID does not exist" containerID="784e9766ec3aebd02560a716704c4d48760309927d4008e2c136d460249ddb98" Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.921048 1425 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:784e9766ec3aebd02560a716704c4d48760309927d4008e2c136d460249ddb98} err="failed to get container status \"784e9766ec3aebd02560a716704c4d48760309927d4008e2c136d460249ddb98\": rpc error: code = NotFound desc = could not find container \"784e9766ec3aebd02560a716704c4d48760309927d4008e2c136d460249ddb98\": container with ID starting with 784e9766ec3aebd02560a716704c4d48760309927d4008e2c136d460249ddb98 not found: ID does not exist" Jan 05 09:14:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:37.925994 1425 kubelet.go:2129] "SyncLoop REMOVE" source="api" pods=[openshift-image-registry/image-registry-64655f6865-dx2q6] Jan 05 09:14:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:38.666244 1425 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=4685a26b-f07f-4b10-8547-dea3d6802723 path="/var/lib/kubelet/pods/4685a26b-f07f-4b10-8547-dea3d6802723/volumes" Jan 05 09:14:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:42.780054 1425 pod_container_manager_linux.go:191] "Failed to delete cgroup paths" cgroupName=[kubepods burstable pod8ac551e8-a959-4212-86a1-2b3ac482586f] err="unable to destroy cgroup paths for cgroup [kubepods burstable pod8ac551e8-a959-4212-86a1-2b3ac482586f] : Timed out while waiting for systemd to remove kubepods-burstable-pod8ac551e8_a959_4212_86a1_2b3ac482586f.slice" Jan 05 09:14:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:44.759989 1425 patch_prober.go:29] interesting pod/router-default-5cf6bbdff9-fkmcn container/router namespace/openshift-ingress: Readiness probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]backend-http ok Jan 05 09:14:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: [+]has-synced ok Jan 05 09:14:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: [-]process-running failed: reason withheld Jan 05 09:14:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: healthz check failed Jan 05 09:14:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:44.760051 1425 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" podUID=66d22b70-08bb-4199-b258-ac4590e9b487 containerName="router" probeResult=failure output="HTTP probe failed with statuscode: 500" Jan 05 09:14:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:54.760139 1425 patch_prober.go:29] interesting pod/router-default-5cf6bbdff9-fkmcn container/router namespace/openshift-ingress: Readiness probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]backend-http ok Jan 05 09:14:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: [+]has-synced ok Jan 05 09:14:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: [-]process-running failed: reason withheld Jan 05 09:14:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: healthz check failed Jan 05 09:14:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:14:54.760197 1425 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" podUID=66d22b70-08bb-4199-b258-ac4590e9b487 containerName="router" probeResult=failure output="HTTP probe failed with statuscode: 500" Jan 05 09:14:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:14:57.632802332Z" level=warning msg="Found defunct process with PID 76161 (haproxy)" Jan 05 09:15:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:04.760575 1425 patch_prober.go:29] interesting pod/router-default-5cf6bbdff9-fkmcn container/router namespace/openshift-ingress: Readiness probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 05 09:15:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: [+]has-synced ok Jan 05 09:15:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: [-]process-running failed: reason withheld Jan 05 09:15:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: healthz check failed Jan 05 09:15:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:04.760638 1425 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" podUID=66d22b70-08bb-4199-b258-ac4590e9b487 containerName="router" probeResult=failure output="HTTP probe failed with statuscode: 500" Jan 05 09:15:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:14.759599 1425 patch_prober.go:29] interesting pod/router-default-5cf6bbdff9-fkmcn container/router namespace/openshift-ingress: Readiness probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 05 09:15:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: [+]has-synced ok Jan 05 09:15:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: [-]process-running failed: reason withheld Jan 05 09:15:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: healthz check failed Jan 05 09:15:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:14.759659 1425 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" podUID=66d22b70-08bb-4199-b258-ac4590e9b487 containerName="router" probeResult=failure output="HTTP probe failed with statuscode: 500" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.727635 1425 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-multus/cni-sysctl-allowlist-ds-tpxqs] Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.727680 1425 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:15:22.727739 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="8ac551e8-a959-4212-86a1-2b3ac482586f" containerName="prometheus-proxy" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.727751 1425 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ac551e8-a959-4212-86a1-2b3ac482586f" containerName="prometheus-proxy" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:15:22.727765 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="b59f4808-02df-4b2e-9694-8cf72a1b5e75" containerName="alertmanager-proxy" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.727774 1425 state_mem.go:107] "Deleted CPUSet assignment" podUID="b59f4808-02df-4b2e-9694-8cf72a1b5e75" containerName="alertmanager-proxy" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:15:22.727789 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="4685a26b-f07f-4b10-8547-dea3d6802723" containerName="registry" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.727798 1425 state_mem.go:107] "Deleted CPUSet assignment" podUID="4685a26b-f07f-4b10-8547-dea3d6802723" containerName="registry" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:15:22.727812 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="89a48689-810c-464e-9ac6-d78977e60793" containerName="check-endpoints" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.727820 1425 state_mem.go:107] "Deleted CPUSet assignment" podUID="89a48689-810c-464e-9ac6-d78977e60793" containerName="check-endpoints" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:15:22.727831 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c" containerName="kube-rbac-proxy-rules" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.727842 1425 state_mem.go:107] "Deleted CPUSet assignment" podUID="72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c" containerName="kube-rbac-proxy-rules" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:15:22.727854 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c" containerName="kube-rbac-proxy-metrics" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.727863 1425 state_mem.go:107] "Deleted CPUSet assignment" podUID="72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c" containerName="kube-rbac-proxy-metrics" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:15:22.727876 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="a3cb294f-6846-4e6b-b42f-434294a1cefd" containerName="kube-multus-additional-cni-plugins" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.727886 1425 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3cb294f-6846-4e6b-b42f-434294a1cefd" containerName="kube-multus-additional-cni-plugins" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:15:22.727896 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="f8e5be3f-9ab7-484c-b61b-6231a8774e24" containerName="prometheus-operator-admission-webhook" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.727905 1425 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8e5be3f-9ab7-484c-b61b-6231a8774e24" containerName="prometheus-operator-admission-webhook" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:15:22.727916 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c" containerName="kube-rbac-proxy" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.727928 1425 state_mem.go:107] "Deleted CPUSet assignment" podUID="72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c" containerName="kube-rbac-proxy" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:15:22.727939 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="8ac551e8-a959-4212-86a1-2b3ac482586f" containerName="config-reloader" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.727948 1425 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ac551e8-a959-4212-86a1-2b3ac482586f" containerName="config-reloader" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:15:22.727957 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c" containerName="oauth-proxy" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.727964 1425 state_mem.go:107] "Deleted CPUSet assignment" podUID="72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c" containerName="oauth-proxy" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:15:22.727973 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="8ac551e8-a959-4212-86a1-2b3ac482586f" containerName="kube-rbac-proxy" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.727980 1425 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ac551e8-a959-4212-86a1-2b3ac482586f" containerName="kube-rbac-proxy" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:15:22.727991 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="8ac551e8-a959-4212-86a1-2b3ac482586f" containerName="thanos-sidecar" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728000 1425 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ac551e8-a959-4212-86a1-2b3ac482586f" containerName="thanos-sidecar" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:15:22.728010 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="a82fd83b-5344-4e7a-91dd-5894f3846e87" containerName="prometheus-adapter" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728017 1425 state_mem.go:107] "Deleted CPUSet assignment" podUID="a82fd83b-5344-4e7a-91dd-5894f3846e87" containerName="prometheus-adapter" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:15:22.728026 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="b59f4808-02df-4b2e-9694-8cf72a1b5e75" containerName="prom-label-proxy" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728034 1425 state_mem.go:107] "Deleted CPUSet assignment" podUID="b59f4808-02df-4b2e-9694-8cf72a1b5e75" containerName="prom-label-proxy" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:15:22.728043 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="8ac551e8-a959-4212-86a1-2b3ac482586f" containerName="kube-rbac-proxy-thanos" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728051 1425 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ac551e8-a959-4212-86a1-2b3ac482586f" containerName="kube-rbac-proxy-thanos" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:15:22.728060 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c" containerName="thanos-query" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728071 1425 state_mem.go:107] "Deleted CPUSet assignment" podUID="72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c" containerName="thanos-query" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:15:22.728080 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="b59f4808-02df-4b2e-9694-8cf72a1b5e75" containerName="alertmanager" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728087 1425 state_mem.go:107] "Deleted CPUSet assignment" podUID="b59f4808-02df-4b2e-9694-8cf72a1b5e75" containerName="alertmanager" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:15:22.728096 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="b59f4808-02df-4b2e-9694-8cf72a1b5e75" containerName="kube-rbac-proxy-metric" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728104 1425 state_mem.go:107] "Deleted CPUSet assignment" podUID="b59f4808-02df-4b2e-9694-8cf72a1b5e75" containerName="kube-rbac-proxy-metric" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:15:22.728114 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="26f1eb05-de00-4fc2-864c-6b97e7f82e96" containerName="migrator" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728123 1425 state_mem.go:107] "Deleted CPUSet assignment" podUID="26f1eb05-de00-4fc2-864c-6b97e7f82e96" containerName="migrator" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:15:22.728134 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="8ac551e8-a959-4212-86a1-2b3ac482586f" containerName="init-config-reloader" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728142 1425 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ac551e8-a959-4212-86a1-2b3ac482586f" containerName="init-config-reloader" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:15:22.728153 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="b59f4808-02df-4b2e-9694-8cf72a1b5e75" containerName="config-reloader" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728161 1425 state_mem.go:107] "Deleted CPUSet assignment" podUID="b59f4808-02df-4b2e-9694-8cf72a1b5e75" containerName="config-reloader" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:15:22.728170 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c" containerName="prom-label-proxy" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728179 1425 state_mem.go:107] "Deleted CPUSet assignment" podUID="72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c" containerName="prom-label-proxy" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:15:22.728190 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="b59f4808-02df-4b2e-9694-8cf72a1b5e75" containerName="kube-rbac-proxy" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728199 1425 state_mem.go:107] "Deleted CPUSet assignment" podUID="b59f4808-02df-4b2e-9694-8cf72a1b5e75" containerName="kube-rbac-proxy" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:15:22.728210 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="8ac551e8-a959-4212-86a1-2b3ac482586f" containerName="prometheus" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728220 1425 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ac551e8-a959-4212-86a1-2b3ac482586f" containerName="prometheus" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728274 1425 memory_manager.go:345] "RemoveStaleState removing state" podUID="a82fd83b-5344-4e7a-91dd-5894f3846e87" containerName="prometheus-adapter" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728286 1425 memory_manager.go:345] "RemoveStaleState removing state" podUID="b59f4808-02df-4b2e-9694-8cf72a1b5e75" containerName="kube-rbac-proxy" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728297 1425 memory_manager.go:345] "RemoveStaleState removing state" podUID="72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c" containerName="thanos-query" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728307 1425 memory_manager.go:345] "RemoveStaleState removing state" podUID="8ac551e8-a959-4212-86a1-2b3ac482586f" containerName="prometheus" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728318 1425 memory_manager.go:345] "RemoveStaleState removing state" podUID="8ac551e8-a959-4212-86a1-2b3ac482586f" containerName="prometheus-proxy" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728329 1425 memory_manager.go:345] "RemoveStaleState removing state" podUID="b59f4808-02df-4b2e-9694-8cf72a1b5e75" containerName="config-reloader" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728340 1425 memory_manager.go:345] "RemoveStaleState removing state" podUID="8ac551e8-a959-4212-86a1-2b3ac482586f" containerName="kube-rbac-proxy-thanos" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728351 1425 memory_manager.go:345] "RemoveStaleState removing state" podUID="72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c" containerName="prom-label-proxy" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728360 1425 memory_manager.go:345] "RemoveStaleState removing state" podUID="b59f4808-02df-4b2e-9694-8cf72a1b5e75" containerName="alertmanager" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728371 1425 memory_manager.go:345] "RemoveStaleState removing state" podUID="b59f4808-02df-4b2e-9694-8cf72a1b5e75" containerName="kube-rbac-proxy-metric" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728381 1425 memory_manager.go:345] "RemoveStaleState removing state" podUID="f8e5be3f-9ab7-484c-b61b-6231a8774e24" containerName="prometheus-operator-admission-webhook" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728411 1425 memory_manager.go:345] "RemoveStaleState removing state" podUID="26f1eb05-de00-4fc2-864c-6b97e7f82e96" containerName="migrator" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728422 1425 memory_manager.go:345] "RemoveStaleState removing state" podUID="a3cb294f-6846-4e6b-b42f-434294a1cefd" containerName="kube-multus-additional-cni-plugins" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728432 1425 memory_manager.go:345] "RemoveStaleState removing state" podUID="8ac551e8-a959-4212-86a1-2b3ac482586f" containerName="kube-rbac-proxy" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728441 1425 memory_manager.go:345] "RemoveStaleState removing state" podUID="b59f4808-02df-4b2e-9694-8cf72a1b5e75" containerName="prom-label-proxy" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728453 1425 memory_manager.go:345] "RemoveStaleState removing state" podUID="72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c" containerName="kube-rbac-proxy" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728464 1425 memory_manager.go:345] "RemoveStaleState removing state" podUID="72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c" containerName="kube-rbac-proxy-rules" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728475 1425 memory_manager.go:345] "RemoveStaleState removing state" podUID="4685a26b-f07f-4b10-8547-dea3d6802723" containerName="registry" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728486 1425 memory_manager.go:345] "RemoveStaleState removing state" podUID="89a48689-810c-464e-9ac6-d78977e60793" containerName="check-endpoints" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728495 1425 memory_manager.go:345] "RemoveStaleState removing state" podUID="72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c" containerName="kube-rbac-proxy-metrics" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728506 1425 memory_manager.go:345] "RemoveStaleState removing state" podUID="8ac551e8-a959-4212-86a1-2b3ac482586f" containerName="config-reloader" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728515 1425 memory_manager.go:345] "RemoveStaleState removing state" podUID="8ac551e8-a959-4212-86a1-2b3ac482586f" containerName="thanos-sidecar" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728525 1425 memory_manager.go:345] "RemoveStaleState removing state" podUID="b59f4808-02df-4b2e-9694-8cf72a1b5e75" containerName="alertmanager-proxy" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.728534 1425 memory_manager.go:345] "RemoveStaleState removing state" podUID="72bf5a0d-4890-43f0-b0f9-b3a54ce7c85c" containerName="oauth-proxy" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.805437 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/02e0075b-c660-4dc5-8cce-00fc4ef2980f-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-tpxqs\" (UID: \"02e0075b-c660-4dc5-8cce-00fc4ef2980f\") " pod="openshift-multus/cni-sysctl-allowlist-ds-tpxqs" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.805486 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/02e0075b-c660-4dc5-8cce-00fc4ef2980f-ready\") pod \"cni-sysctl-allowlist-ds-tpxqs\" (UID: \"02e0075b-c660-4dc5-8cce-00fc4ef2980f\") " pod="openshift-multus/cni-sysctl-allowlist-ds-tpxqs" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.805518 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fwhf\" (UniqueName: \"kubernetes.io/projected/02e0075b-c660-4dc5-8cce-00fc4ef2980f-kube-api-access-7fwhf\") pod \"cni-sysctl-allowlist-ds-tpxqs\" (UID: \"02e0075b-c660-4dc5-8cce-00fc4ef2980f\") " pod="openshift-multus/cni-sysctl-allowlist-ds-tpxqs" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.805565 1425 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/02e0075b-c660-4dc5-8cce-00fc4ef2980f-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-tpxqs\" (UID: \"02e0075b-c660-4dc5-8cce-00fc4ef2980f\") " pod="openshift-multus/cni-sysctl-allowlist-ds-tpxqs" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.905905 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/02e0075b-c660-4dc5-8cce-00fc4ef2980f-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-tpxqs\" (UID: \"02e0075b-c660-4dc5-8cce-00fc4ef2980f\") " pod="openshift-multus/cni-sysctl-allowlist-ds-tpxqs" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.905957 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/02e0075b-c660-4dc5-8cce-00fc4ef2980f-ready\") pod \"cni-sysctl-allowlist-ds-tpxqs\" (UID: \"02e0075b-c660-4dc5-8cce-00fc4ef2980f\") " pod="openshift-multus/cni-sysctl-allowlist-ds-tpxqs" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.905989 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-7fwhf\" (UniqueName: \"kubernetes.io/projected/02e0075b-c660-4dc5-8cce-00fc4ef2980f-kube-api-access-7fwhf\") pod \"cni-sysctl-allowlist-ds-tpxqs\" (UID: \"02e0075b-c660-4dc5-8cce-00fc4ef2980f\") " pod="openshift-multus/cni-sysctl-allowlist-ds-tpxqs" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.906025 1425 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/02e0075b-c660-4dc5-8cce-00fc4ef2980f-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-tpxqs\" (UID: \"02e0075b-c660-4dc5-8cce-00fc4ef2980f\") " pod="openshift-multus/cni-sysctl-allowlist-ds-tpxqs" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.906108 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/02e0075b-c660-4dc5-8cce-00fc4ef2980f-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-tpxqs\" (UID: \"02e0075b-c660-4dc5-8cce-00fc4ef2980f\") " pod="openshift-multus/cni-sysctl-allowlist-ds-tpxqs" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.906480 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/02e0075b-c660-4dc5-8cce-00fc4ef2980f-ready\") pod \"cni-sysctl-allowlist-ds-tpxqs\" (UID: \"02e0075b-c660-4dc5-8cce-00fc4ef2980f\") " pod="openshift-multus/cni-sysctl-allowlist-ds-tpxqs" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.906691 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/02e0075b-c660-4dc5-8cce-00fc4ef2980f-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-tpxqs\" (UID: \"02e0075b-c660-4dc5-8cce-00fc4ef2980f\") " pod="openshift-multus/cni-sysctl-allowlist-ds-tpxqs" Jan 05 09:15:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:22.925785 1425 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fwhf\" (UniqueName: \"kubernetes.io/projected/02e0075b-c660-4dc5-8cce-00fc4ef2980f-kube-api-access-7fwhf\") pod \"cni-sysctl-allowlist-ds-tpxqs\" (UID: \"02e0075b-c660-4dc5-8cce-00fc4ef2980f\") " pod="openshift-multus/cni-sysctl-allowlist-ds-tpxqs" Jan 05 09:15:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:23.047640 1425 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-tpxqs" Jan 05 09:15:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:23.048087082Z" level=info msg="Running pod sandbox: openshift-multus/cni-sysctl-allowlist-ds-tpxqs/POD" id=18aa517f-c711-4016-a9c5-f218f2d07279 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:15:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:23.048146252Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:15:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:23.078037608Z" level=info msg="Got pod network &{Name:cni-sysctl-allowlist-ds-tpxqs Namespace:openshift-multus ID:1e1cece1b89c1fd30700275a80eef602cac31d5d3e2d8e7903561c25eafb4a33 UID:02e0075b-c660-4dc5-8cce-00fc4ef2980f NetNS:/var/run/netns/2e7edc50-42f9-414e-9cc3-147bbac4956f Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:15:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:23.078073863Z" level=info msg="Adding pod openshift-multus_cni-sysctl-allowlist-ds-tpxqs to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:24.558140269Z" level=info msg="Stopped container 8c886b3e9fbca71166a4d01928168d286331774947ddef0326c331f3624f5f6a: openshift-ingress/router-default-5cf6bbdff9-fkmcn/router" id=9c26b5a4-93b8-4eb7-ba4a-6d69c5d1f52d name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:24.558737240Z" level=info msg="Stopping pod sandbox: 0412a26df3b1c530016b6935b36a365c8ab277d3b739d57e9cec85e5975d314c" id=ce383059-8b0b-4554-8ddc-6b4c9894cc36 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:24.558953848Z" level=info msg="Got pod network &{Name:router-default-5cf6bbdff9-fkmcn Namespace:openshift-ingress ID:0412a26df3b1c530016b6935b36a365c8ab277d3b739d57e9cec85e5975d314c UID:66d22b70-08bb-4199-b258-ac4590e9b487 NetNS:/var/run/netns/c442ffc5-f1f5-46e0-870e-1ec985722737 Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:24.559076057Z" level=info msg="Deleting pod openshift-ingress_router-default-5cf6bbdff9-fkmcn from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: 2023-01-05T09:15:24Z [verbose] Del: openshift-ingress:router-default-5cf6bbdff9-fkmcn:66d22b70-08bb-4199-b258-ac4590e9b487:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:24.704704509Z" level=info msg="Stopped pod sandbox: 0412a26df3b1c530016b6935b36a365c8ab277d3b739d57e9cec85e5975d314c" id=ce383059-8b0b-4554-8ddc-6b4c9894cc36 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:24.817576 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/66d22b70-08bb-4199-b258-ac4590e9b487-stats-auth\") pod \"66d22b70-08bb-4199-b258-ac4590e9b487\" (UID: \"66d22b70-08bb-4199-b258-ac4590e9b487\") " Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:24.817638 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/66d22b70-08bb-4199-b258-ac4590e9b487-default-certificate\") pod \"66d22b70-08bb-4199-b258-ac4590e9b487\" (UID: \"66d22b70-08bb-4199-b258-ac4590e9b487\") " Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:24.817673 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66d22b70-08bb-4199-b258-ac4590e9b487-service-ca-bundle\") pod \"66d22b70-08bb-4199-b258-ac4590e9b487\" (UID: \"66d22b70-08bb-4199-b258-ac4590e9b487\") " Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:24.817708 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzl7h\" (UniqueName: \"kubernetes.io/projected/66d22b70-08bb-4199-b258-ac4590e9b487-kube-api-access-vzl7h\") pod \"66d22b70-08bb-4199-b258-ac4590e9b487\" (UID: \"66d22b70-08bb-4199-b258-ac4590e9b487\") " Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:24.817743 1425 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/66d22b70-08bb-4199-b258-ac4590e9b487-metrics-certs\") pod \"66d22b70-08bb-4199-b258-ac4590e9b487\" (UID: \"66d22b70-08bb-4199-b258-ac4590e9b487\") " Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:15:24.818636 1425 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/66d22b70-08bb-4199-b258-ac4590e9b487/volumes/kubernetes.io~configmap/service-ca-bundle: clearQuota called, but quotas disabled Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:24.818845 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66d22b70-08bb-4199-b258-ac4590e9b487-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "66d22b70-08bb-4199-b258-ac4590e9b487" (UID: "66d22b70-08bb-4199-b258-ac4590e9b487"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:24.826004 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66d22b70-08bb-4199-b258-ac4590e9b487-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "66d22b70-08bb-4199-b258-ac4590e9b487" (UID: "66d22b70-08bb-4199-b258-ac4590e9b487"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:24.827978 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66d22b70-08bb-4199-b258-ac4590e9b487-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "66d22b70-08bb-4199-b258-ac4590e9b487" (UID: "66d22b70-08bb-4199-b258-ac4590e9b487"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:24.827984 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66d22b70-08bb-4199-b258-ac4590e9b487-kube-api-access-vzl7h" (OuterVolumeSpecName: "kube-api-access-vzl7h") pod "66d22b70-08bb-4199-b258-ac4590e9b487" (UID: "66d22b70-08bb-4199-b258-ac4590e9b487"). InnerVolumeSpecName "kube-api-access-vzl7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:24.828359 1425 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66d22b70-08bb-4199-b258-ac4590e9b487-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "66d22b70-08bb-4199-b258-ac4590e9b487" (UID: "66d22b70-08bb-4199-b258-ac4590e9b487"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:24.895632 1425 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-multus/cni-sysctl-allowlist-ds-tpxqs] Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: 2023-01-05T09:15:24Z [verbose] Add: openshift-multus:cni-sysctl-allowlist-ds-tpxqs:02e0075b-c660-4dc5-8cce-00fc4ef2980f:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/2e7edc50-42f9-414e-9cc3-147bbac4956f"}],"ips":[{"version":"4","interface":0,"address":"10.129.2.16/23","gateway":"10.129.2.1"}],"dns":{}} Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: I0105 09:15:24.876527 79934 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-multus", Name:"cni-sysctl-allowlist-ds-tpxqs", UID:"02e0075b-c660-4dc5-8cce-00fc4ef2980f", APIVersion:"v1", ResourceVersion:"78520", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.129.2.16/23] from openshift-sdn Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:24.896196586Z" level=info msg="Got pod network &{Name:cni-sysctl-allowlist-ds-tpxqs Namespace:openshift-multus ID:1e1cece1b89c1fd30700275a80eef602cac31d5d3e2d8e7903561c25eafb4a33 UID:02e0075b-c660-4dc5-8cce-00fc4ef2980f NetNS:/var/run/netns/2e7edc50-42f9-414e-9cc3-147bbac4956f Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:24.896334632Z" level=info msg="Checking pod openshift-multus_cni-sysctl-allowlist-ds-tpxqs for CNI network multus-cni-network (type=multus)" Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: W0105 09:15:24.898128 1425 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02e0075b_c660_4dc5_8cce_00fc4ef2980f.slice/crio-1e1cece1b89c1fd30700275a80eef602cac31d5d3e2d8e7903561c25eafb4a33.scope WatchSource:0}: Error finding container 1e1cece1b89c1fd30700275a80eef602cac31d5d3e2d8e7903561c25eafb4a33: Status 404 returned error can't find the container with id 1e1cece1b89c1fd30700275a80eef602cac31d5d3e2d8e7903561c25eafb4a33 Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:24.902469652Z" level=info msg="Ran pod sandbox 1e1cece1b89c1fd30700275a80eef602cac31d5d3e2d8e7903561c25eafb4a33 with infra container: openshift-multus/cni-sysctl-allowlist-ds-tpxqs/POD" id=18aa517f-c711-4016-a9c5-f218f2d07279 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:24.903198121Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=bf348598-c5db-4b5d-8c1b-9741c7346aba name=/runtime.v1.ImageService/ImageStatus Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:24.903433423Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=bf348598-c5db-4b5d-8c1b-9741c7346aba name=/runtime.v1.ImageService/ImageStatus Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:24.904048474Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=babda4ab-51f4-40db-a55e-afe1eeab9c2c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:24.904236518Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=babda4ab-51f4-40db-a55e-afe1eeab9c2c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:24.904918843Z" level=info msg="Creating container: openshift-multus/cni-sysctl-allowlist-ds-tpxqs/kube-multus-additional-cni-plugins" id=5f53fdd5-83ea-4882-9c74-f67e6a6f3678 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:24.905032332Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:24.918318 1425 reconciler.go:399] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/66d22b70-08bb-4199-b258-ac4590e9b487-default-certificate\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:24.918349 1425 reconciler.go:399] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66d22b70-08bb-4199-b258-ac4590e9b487-service-ca-bundle\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:24.918367 1425 reconciler.go:399] "Volume detached for volume \"kube-api-access-vzl7h\" (UniqueName: \"kubernetes.io/projected/66d22b70-08bb-4199-b258-ac4590e9b487-kube-api-access-vzl7h\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:24.918384 1425 reconciler.go:399] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/66d22b70-08bb-4199-b258-ac4590e9b487-metrics-certs\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:24.918417 1425 reconciler.go:399] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/66d22b70-08bb-4199-b258-ac4590e9b487-stats-auth\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:24.979449 1425 generic.go:296] "Generic (PLEG): container finished" podID=66d22b70-08bb-4199-b258-ac4590e9b487 containerID="8c886b3e9fbca71166a4d01928168d286331774947ddef0326c331f3624f5f6a" exitCode=0 Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:24.979524 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" event=&{ID:66d22b70-08bb-4199-b258-ac4590e9b487 Type:ContainerDied Data:8c886b3e9fbca71166a4d01928168d286331774947ddef0326c331f3624f5f6a} Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:24.979563 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5cf6bbdff9-fkmcn" event=&{ID:66d22b70-08bb-4199-b258-ac4590e9b487 Type:ContainerDied Data:0412a26df3b1c530016b6935b36a365c8ab277d3b739d57e9cec85e5975d314c} Jan 05 09:15:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:24.979586 1425 scope.go:115] "RemoveContainer" containerID="8c886b3e9fbca71166a4d01928168d286331774947ddef0326c331f3624f5f6a" Jan 05 09:15:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:25.015021365Z" level=info msg="Removing container: 8c886b3e9fbca71166a4d01928168d286331774947ddef0326c331f3624f5f6a" id=8a256abb-5b24-43eb-9578-65aafeeb314d name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:15:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:25.024100 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-tpxqs" event=&{ID:02e0075b-c660-4dc5-8cce-00fc4ef2980f Type:ContainerStarted Data:1e1cece1b89c1fd30700275a80eef602cac31d5d3e2d8e7903561c25eafb4a33} Jan 05 09:15:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:25.027575 1425 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-ingress/router-default-5cf6bbdff9-fkmcn] Jan 05 09:15:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:25.035848 1425 kubelet.go:2129] "SyncLoop REMOVE" source="api" pods=[openshift-ingress/router-default-5cf6bbdff9-fkmcn] Jan 05 09:15:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:25.060653320Z" level=info msg="Removed container 8c886b3e9fbca71166a4d01928168d286331774947ddef0326c331f3624f5f6a: openshift-ingress/router-default-5cf6bbdff9-fkmcn/router" id=8a256abb-5b24-43eb-9578-65aafeeb314d name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:15:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:25.060872 1425 scope.go:115] "RemoveContainer" containerID="8c886b3e9fbca71166a4d01928168d286331774947ddef0326c331f3624f5f6a" Jan 05 09:15:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: E0105 09:15:25.061206 1425 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c886b3e9fbca71166a4d01928168d286331774947ddef0326c331f3624f5f6a\": container with ID starting with 8c886b3e9fbca71166a4d01928168d286331774947ddef0326c331f3624f5f6a not found: ID does not exist" containerID="8c886b3e9fbca71166a4d01928168d286331774947ddef0326c331f3624f5f6a" Jan 05 09:15:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:25.061245 1425 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:8c886b3e9fbca71166a4d01928168d286331774947ddef0326c331f3624f5f6a} err="failed to get container status \"8c886b3e9fbca71166a4d01928168d286331774947ddef0326c331f3624f5f6a\": rpc error: code = NotFound desc = could not find container \"8c886b3e9fbca71166a4d01928168d286331774947ddef0326c331f3624f5f6a\": container with ID starting with 8c886b3e9fbca71166a4d01928168d286331774947ddef0326c331f3624f5f6a not found: ID does not exist" Jan 05 09:15:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:25.121533388Z" level=info msg="Created container 302e2f3529c590dfc8f53fc352ffb3636352b16ca70ce41c7c522e56153ffe51: openshift-multus/cni-sysctl-allowlist-ds-tpxqs/kube-multus-additional-cni-plugins" id=5f53fdd5-83ea-4882-9c74-f67e6a6f3678 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:15:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:25.122203462Z" level=info msg="Starting container: 302e2f3529c590dfc8f53fc352ffb3636352b16ca70ce41c7c522e56153ffe51" id=cd80d783-a2cd-42a3-bfe9-567ea8d2b1c8 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:15:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:25.135096836Z" level=info msg="Started container" PID=80118 containerID=302e2f3529c590dfc8f53fc352ffb3636352b16ca70ce41c7c522e56153ffe51 description=openshift-multus/cni-sysctl-allowlist-ds-tpxqs/kube-multus-additional-cni-plugins id=cd80d783-a2cd-42a3-bfe9-567ea8d2b1c8 name=/runtime.v1.RuntimeService/StartContainer sandboxID=1e1cece1b89c1fd30700275a80eef602cac31d5d3e2d8e7903561c25eafb4a33 Jan 05 09:15:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:26.027764 1425 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-tpxqs" event=&{ID:02e0075b-c660-4dc5-8cce-00fc4ef2980f Type:ContainerStarted Data:302e2f3529c590dfc8f53fc352ffb3636352b16ca70ce41c7c522e56153ffe51} Jan 05 09:15:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:26.028684 1425 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-multus/cni-sysctl-allowlist-ds-tpxqs" Jan 05 09:15:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:26.119123 1425 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-multus/cni-sysctl-allowlist-ds-tpxqs" Jan 05 09:15:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:26.664554450Z" level=info msg="Stopping pod sandbox: 0412a26df3b1c530016b6935b36a365c8ab277d3b739d57e9cec85e5975d314c" id=00a8f827-5959-4b07-8697-6c326df2c1f6 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:15:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:26.664598604Z" level=info msg="Stopped pod sandbox (already stopped): 0412a26df3b1c530016b6935b36a365c8ab277d3b739d57e9cec85e5975d314c" id=00a8f827-5959-4b07-8697-6c326df2c1f6 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:15:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:26.666201 1425 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=66d22b70-08bb-4199-b258-ac4590e9b487 path="/var/lib/kubelet/pods/66d22b70-08bb-4199-b258-ac4590e9b487/volumes" Jan 05 09:15:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:26.748606 1425 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-multus/cni-sysctl-allowlist-ds-tpxqs] Jan 05 09:15:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:28.031964 1425 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-multus/cni-sysctl-allowlist-ds-tpxqs" podUID=02e0075b-c660-4dc5-8cce-00fc4ef2980f containerName="kube-multus-additional-cni-plugins" containerID="cri-o://302e2f3529c590dfc8f53fc352ffb3636352b16ca70ce41c7c522e56153ffe51" gracePeriod=30 Jan 05 09:15:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:28.032584372Z" level=info msg="Stopping container: 302e2f3529c590dfc8f53fc352ffb3636352b16ca70ce41c7c522e56153ffe51 (timeout: 30s)" id=f0fff3e3-e0a3-4db6-9a9d-ec6276715df3 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:15:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:36.629514 1425 dynamic_cafile_content.go:211] "Failed to remove file watch, it may have been deleted" file="/etc/kubernetes/kubelet-ca.crt" err="can't remove non-existent inotify watch for: /etc/kubernetes/kubelet-ca.crt" Jan 05 09:15:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:36.739388537Z" level=info msg="Stopping pod sandbox: 0412a26df3b1c530016b6935b36a365c8ab277d3b739d57e9cec85e5975d314c" id=8bcc0d50-de8c-4d88-9d1d-59ea4d99e7de name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:15:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:36.739457167Z" level=info msg="Stopped pod sandbox (already stopped): 0412a26df3b1c530016b6935b36a365c8ab277d3b739d57e9cec85e5975d314c" id=8bcc0d50-de8c-4d88-9d1d-59ea4d99e7de name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:15:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:36.739691716Z" level=info msg="Removing pod sandbox: 0412a26df3b1c530016b6935b36a365c8ab277d3b739d57e9cec85e5975d314c" id=05a81438-d8c1-44db-a7f3-f367171a496c name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:15:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:36.755595274Z" level=info msg="Removed pod sandbox: 0412a26df3b1c530016b6935b36a365c8ab277d3b739d57e9cec85e5975d314c" id=05a81438-d8c1-44db-a7f3-f367171a496c name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:15:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:36.765964099Z" level=info msg="Stopping pod sandbox: f855fd38c50954c79bc8f0f3e034b1d0f7293735e9dbc194b9e271cfb674c01c" id=a3291521-0b0d-40f1-b3b5-0b2ee3857ae2 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:15:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:36.765999678Z" level=info msg="Stopped pod sandbox (already stopped): f855fd38c50954c79bc8f0f3e034b1d0f7293735e9dbc194b9e271cfb674c01c" id=a3291521-0b0d-40f1-b3b5-0b2ee3857ae2 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:15:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:36.779574062Z" level=info msg="Removing pod sandbox: f855fd38c50954c79bc8f0f3e034b1d0f7293735e9dbc194b9e271cfb674c01c" id=f3b3dc0f-2ed8-4e94-9fa2-e581619cb93e name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:15:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:36.795099399Z" level=info msg="Removed pod sandbox: f855fd38c50954c79bc8f0f3e034b1d0f7293735e9dbc194b9e271cfb674c01c" id=f3b3dc0f-2ed8-4e94-9fa2-e581619cb93e name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:15:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Stopping Kubernetes Kubelet... Jan 05 09:15:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1425]: I0105 09:15:38.658302 1425 dynamic_cafile_content.go:171] "Shutting down controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 05 09:15:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:39.007222756Z" level=warning msg="Stopping container 302e2f3529c590dfc8f53fc352ffb3636352b16ca70ce41c7c522e56153ffe51 with stop signal timed out: context canceled" id=f0fff3e3-e0a3-4db6-9a9d-ec6276715df3 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:15:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: kubelet.service: Succeeded. Jan 05 09:15:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Stopped Kubernetes Kubelet. Jan 05 09:15:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: kubelet.service: Consumed 3min 41.549s CPU time Jan 05 09:15:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Stopping Container Runtime Interface for OCI (CRI-O)... Jan 05 09:15:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1396]: time="2023-01-05 09:15:58.653648081Z" level=error msg="Failed to update container state for e510b9127a187a68744a288407216f7bbd76dfd74cc9d05d826a3e7cc3ecba24: `/usr/bin/runc --root /run/runc --systemd-cgroup state e510b9127a187a68744a288407216f7bbd76dfd74cc9d05d826a3e7cc3ecba24` failed: : signal: terminated" Jan 05 09:15:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: crio.service: Succeeded. Jan 05 09:15:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Stopped Container Runtime Interface for OCI (CRI-O). Jan 05 09:15:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: crio.service: Consumed 3min 15.349s CPU time -- Reboot -- Jan 05 17:16:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Starting Container Runtime Interface for OCI (CRI-O)... Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.096144403Z" level=info msg="Starting CRI-O, version: 1.25.1-5.rhaos4.12.git6005903.el8, git: unknown(clean)" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.096520088Z" level=info msg="Node configuration value for hugetlb cgroup is true" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.096533741Z" level=info msg="Node configuration value for pid cgroup is true" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.096608980Z" level=info msg="Node configuration value for memoryswap cgroup is true" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.096617742Z" level=info msg="Node configuration value for cgroup v2 is false" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.104552409Z" level=info msg="Node configuration value for systemd CollectMode is true" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.110764039Z" level=info msg="Node configuration value for systemd AllowedCPUs is true" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.113677443Z" level=info msg="Using default capabilities: CAP_CHOWN, CAP_DAC_OVERRIDE, CAP_FSETID, CAP_FOWNER, CAP_SETGID, CAP_SETUID, CAP_SETPCAP, CAP_NET_BIND_SERVICE, CAP_KILL" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.184584980Z" level=info msg="Checkpoint/restore support disabled" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.184750958Z" level=info msg="Using seccomp default profile when unspecified: true" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.184817140Z" level=info msg="Using the internal default seccomp profile" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.184868991Z" level=info msg="AppArmor is disabled by the system or at CRI-O build-time" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.184879437Z" level=info msg="No blockio config file specified, blockio not configured" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.184887414Z" level=info msg="RDT not available in the host system" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.190337334Z" level=info msg="Conmon does support the --sync option" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.190353239Z" level=info msg="Conmon does support the --log-global-size-max option" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.193938605Z" level=info msg="Conmon does support the --sync option" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.193954593Z" level=info msg="Conmon does support the --log-global-size-max option" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.310543600Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.310573790Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.365956034Z" level=warning msg="Could not restore sandbox 77dc8fbfc80728a344c03166326f04c57592cd6f9e2adc95515af9c2147cc500: failed to Statfs \"/var/run/netns/18bb68ea-5b4c-4a61-b216-d8541d4c2c18\": no such file or directory" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.376326535Z" level=warning msg="Deleting all containers under sandbox 77dc8fbfc80728a344c03166326f04c57592cd6f9e2adc95515af9c2147cc500 since it could not be restored" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.402973174Z" level=warning msg="Could not restore sandbox f160f14931d6ef807fb1308f5ac27f062e39c5cad361cedf59e27a46a8b3bd00: failed to Statfs \"/var/run/netns/3bb4165f-22ae-4df5-80c8-98abe16fe8f3\": no such file or directory" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.411161713Z" level=warning msg="Deleting all containers under sandbox f160f14931d6ef807fb1308f5ac27f062e39c5cad361cedf59e27a46a8b3bd00 since it could not be restored" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.424924785Z" level=warning msg="Could not restore sandbox 3e1262384ce877333f952ac7ba12d68b4133181f637b84095f42300e1641af9e: failed to Statfs \"/var/run/netns/91970716-b3de-4e92-b724-0e63fbf7eb3a\": no such file or directory" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.434119580Z" level=warning msg="Deleting all containers under sandbox 3e1262384ce877333f952ac7ba12d68b4133181f637b84095f42300e1641af9e since it could not be restored" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.442893546Z" level=warning msg="Could not restore sandbox 81b981c00eed3ce8e7b961151a721e4be9ef2eef83f4433fad233b19e63776bf: failed to Statfs \"/var/run/netns/7b5530ed-9b90-41d9-b3c4-0a53de335ef7\": no such file or directory" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.450955835Z" level=warning msg="Deleting all containers under sandbox 81b981c00eed3ce8e7b961151a721e4be9ef2eef83f4433fad233b19e63776bf since it could not be restored" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.468659327Z" level=warning msg="Could not restore sandbox d611273f3568d300e64d6a51cb6de91bcf31c3abd27ef78b4fed9dc896457467: failed to Statfs \"/var/run/netns/70366f39-5174-4b35-bd1c-e8cef54fd9ce\": no such file or directory" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.475414538Z" level=warning msg="Deleting all containers under sandbox d611273f3568d300e64d6a51cb6de91bcf31c3abd27ef78b4fed9dc896457467 since it could not be restored" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.503806149Z" level=warning msg="Could not restore sandbox 47c1514985f9106f9edaa2acbae833a41c7f065a534af106d4e033a59d66e2fb: failed to Statfs \"/var/run/netns/db7d401d-1bdc-48f0-9c55-2da49d3a8634\": no such file or directory" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.509626495Z" level=warning msg="Deleting all containers under sandbox 47c1514985f9106f9edaa2acbae833a41c7f065a534af106d4e033a59d66e2fb since it could not be restored" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.519534511Z" level=warning msg="Could not restore sandbox 0b984f430b21595a59efe2294b0b9fce6559367942c6743e20dbab9826dae909: failed to Statfs \"/var/run/netns/af101b6b-a23a-4a11-9eee-7530bdb318d3\": no such file or directory" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.526573816Z" level=warning msg="Deleting all containers under sandbox 0b984f430b21595a59efe2294b0b9fce6559367942c6743e20dbab9826dae909 since it could not be restored" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.580972114Z" level=warning msg="Could not restore sandbox 1e1cece1b89c1fd30700275a80eef602cac31d5d3e2d8e7903561c25eafb4a33: failed to Statfs \"/var/run/netns/2e7edc50-42f9-414e-9cc3-147bbac4956f\": no such file or directory" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.586854651Z" level=warning msg="Deleting all containers under sandbox 1e1cece1b89c1fd30700275a80eef602cac31d5d3e2d8e7903561c25eafb4a33 since it could not be restored" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.592814715Z" level=warning msg="Could not restore sandbox 27b0c41330330b95f91c642446af55e683d47ee10c323ad18406b8c3192efb02: failed to Statfs \"/var/run/netns/238ff1a5-0dfe-4dec-97b0-e24dba3ca3d3\": no such file or directory" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.601312870Z" level=warning msg="Deleting all containers under sandbox 27b0c41330330b95f91c642446af55e683d47ee10c323ad18406b8c3192efb02 since it could not be restored" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.614346581Z" level=warning msg="Could not restore sandbox 7114c6eed49130178fa2f1d156de41d15d9a79d32874a225c7b87798956be18c: failed to Statfs \"/var/run/netns/0dd6b95a-b296-4ff6-b277-2d124d02fb15\": no such file or directory" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.620689534Z" level=warning msg="Deleting all containers under sandbox 7114c6eed49130178fa2f1d156de41d15d9a79d32874a225c7b87798956be18c since it could not be restored" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.639459965Z" level=warning msg="Could not restore sandbox 2cd51e2f6a1cbcdeb26d92c54070ab2a85b957f76038eb175b75e3f2ae639ed7: failed to Statfs \"/var/run/netns/02fd176f-a3ea-4bb8-b8f0-dbf7eac0bad2\": no such file or directory" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.644413294Z" level=warning msg="Deleting all containers under sandbox 2cd51e2f6a1cbcdeb26d92c54070ab2a85b957f76038eb175b75e3f2ae639ed7 since it could not be restored" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.657253593Z" level=warning msg="Could not restore sandbox 5fc6d6ddb9ba864a50f6934237ca81ca5c56501c56049bb7190d5ac76187ea49: failed to Statfs \"/var/run/netns/444697ac-5441-4dc1-82e2-47eb57447726\": no such file or directory" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.666328712Z" level=warning msg="Deleting all containers under sandbox 5fc6d6ddb9ba864a50f6934237ca81ca5c56501c56049bb7190d5ac76187ea49 since it could not be restored" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.677242751Z" level=warning msg="Could not restore sandbox e1cf6641b5c3c13478f683a4318f93d2bf7b7d9376939506acf8a588476f10eb: failed to Statfs \"/var/run/netns/ecce2622-252b-4599-9586-1dab87f71a92\": no such file or directory" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.683358903Z" level=warning msg="Deleting all containers under sandbox e1cf6641b5c3c13478f683a4318f93d2bf7b7d9376939506acf8a588476f10eb since it could not be restored" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.688971025Z" level=warning msg="Could not restore sandbox 63ad0ecbd8ba6a412b2c14895d6a13d0ead1d3ee31d0f15597f1df586d1d96e9: failed to Statfs \"/var/run/netns/0d57e9b1-c13a-46f4-8985-1f615d742ed6\": no such file or directory" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.694446063Z" level=warning msg="Deleting all containers under sandbox 63ad0ecbd8ba6a412b2c14895d6a13d0ead1d3ee31d0f15597f1df586d1d96e9 since it could not be restored" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.702060757Z" level=warning msg="Error encountered when checking whether cri-o should wipe containers: open /var/run/crio/version: no such file or directory" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.702172452Z" level=info msg="cleanup sandbox network" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.703087130Z" level=info msg="Successfully cleaned up network for pod 7114c6eed49130178fa2f1d156de41d15d9a79d32874a225c7b87798956be18c" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.703111345Z" level=info msg="cleanup sandbox network" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.703121484Z" level=info msg="Successfully cleaned up network for pod 27b0c41330330b95f91c642446af55e683d47ee10c323ad18406b8c3192efb02" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.703131529Z" level=info msg="cleanup sandbox network" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.703139835Z" level=info msg="Successfully cleaned up network for pod 47c1514985f9106f9edaa2acbae833a41c7f065a534af106d4e033a59d66e2fb" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.703149204Z" level=info msg="cleanup sandbox network" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.703158275Z" level=info msg="Successfully cleaned up network for pod 77dc8fbfc80728a344c03166326f04c57592cd6f9e2adc95515af9c2147cc500" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.703167996Z" level=info msg="cleanup sandbox network" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.704009718Z" level=info msg="Got pod network &{Name:dns-default-2pfzf Namespace:openshift-dns ID:81b981c00eed3ce8e7b961151a721e4be9ef2eef83f4433fad233b19e63776bf UID:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.704161451Z" level=info msg="Deleting pod openshift-dns_dns-default-2pfzf from CNI network \"multus-cni-network\" (type=multus)" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:21.706150230Z" level=info msg="Serving metrics on :9537 via HTTP" Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Started Container Runtime Interface for OCI (CRI-O). Jan 05 17:16:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Starting Kubernetes Kubelet... Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: Flag --container-runtime has been deprecated, will be removed in 1.27 as the only valid value is 'remote' Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: Flag --cloud-provider has been deprecated, will be removed in 1.25 or later, in favor of removing cloud provider code from Kubelet. Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: Flag --provider-id has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.27. Image garbage collector will get sandbox image information from CRI. Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.078929 2171 server.go:200] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: Flag --container-runtime has been deprecated, will be removed in 1.27 as the only valid value is 'remote' Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: Flag --cloud-provider has been deprecated, will be removed in 1.25 or later, in favor of removing cloud provider code from Kubelet. Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: Flag --provider-id has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.27. Image garbage collector will get sandbox image information from CRI. Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081761 2171 flags.go:64] FLAG: --add-dir-header="false" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081779 2171 flags.go:64] FLAG: --address="0.0.0.0" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081791 2171 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081803 2171 flags.go:64] FLAG: --alsologtostderr="false" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081809 2171 flags.go:64] FLAG: --anonymous-auth="true" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081816 2171 flags.go:64] FLAG: --application-metrics-count-limit="100" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081821 2171 flags.go:64] FLAG: --authentication-token-webhook="false" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081827 2171 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081833 2171 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081840 2171 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081845 2171 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081850 2171 flags.go:64] FLAG: --azure-container-registry-config="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081856 2171 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081861 2171 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081867 2171 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081872 2171 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081877 2171 flags.go:64] FLAG: --cgroup-root="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081882 2171 flags.go:64] FLAG: --cgroups-per-qos="true" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081888 2171 flags.go:64] FLAG: --client-ca-file="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081892 2171 flags.go:64] FLAG: --cloud-config="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081897 2171 flags.go:64] FLAG: --cloud-provider="external" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081902 2171 flags.go:64] FLAG: --cluster-dns="[]" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081916 2171 flags.go:64] FLAG: --cluster-domain="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081922 2171 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081927 2171 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081933 2171 flags.go:64] FLAG: --container-log-max-files="5" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081940 2171 flags.go:64] FLAG: --container-log-max-size="10Mi" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081945 2171 flags.go:64] FLAG: --container-runtime="remote" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081950 2171 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081955 2171 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081961 2171 flags.go:64] FLAG: --containerd-namespace="k8s.io" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081968 2171 flags.go:64] FLAG: --contention-profiling="false" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081973 2171 flags.go:64] FLAG: --cpu-cfs-quota="true" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.081979 2171 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082004 2171 flags.go:64] FLAG: --cpu-manager-policy="none" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082010 2171 flags.go:64] FLAG: --cpu-manager-policy-options="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082017 2171 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082023 2171 flags.go:64] FLAG: --enable-controller-attach-detach="true" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082028 2171 flags.go:64] FLAG: --enable-debugging-handlers="true" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082033 2171 flags.go:64] FLAG: --enable-load-reader="false" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082039 2171 flags.go:64] FLAG: --enable-server="true" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082049 2171 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082060 2171 flags.go:64] FLAG: --event-burst="10" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082065 2171 flags.go:64] FLAG: --event-qps="5" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082070 2171 flags.go:64] FLAG: --event-storage-age-limit="default=0" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082076 2171 flags.go:64] FLAG: --event-storage-event-limit="default=0" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082081 2171 flags.go:64] FLAG: --eviction-hard="imagefs.available<15%,memory.available<100Mi,nodefs.available<10%,nodefs.inodesFree<5%" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082096 2171 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082102 2171 flags.go:64] FLAG: --eviction-minimum-reclaim="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082108 2171 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082113 2171 flags.go:64] FLAG: --eviction-soft="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082118 2171 flags.go:64] FLAG: --eviction-soft-grace-period="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082124 2171 flags.go:64] FLAG: --exit-on-lock-contention="false" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082129 2171 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082134 2171 flags.go:64] FLAG: --experimental-mounter-path="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082139 2171 flags.go:64] FLAG: --fail-swap-on="true" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082145 2171 flags.go:64] FLAG: --feature-gates="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082151 2171 flags.go:64] FLAG: --file-check-frequency="20s" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082157 2171 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082162 2171 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082172 2171 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082178 2171 flags.go:64] FLAG: --healthz-port="10248" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082183 2171 flags.go:64] FLAG: --help="false" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082189 2171 flags.go:64] FLAG: --hostname-override="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082194 2171 flags.go:64] FLAG: --housekeeping-interval="10s" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082200 2171 flags.go:64] FLAG: --http-check-frequency="20s" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082205 2171 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082210 2171 flags.go:64] FLAG: --image-credential-provider-config="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082215 2171 flags.go:64] FLAG: --image-gc-high-threshold="85" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082220 2171 flags.go:64] FLAG: --image-gc-low-threshold="80" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082225 2171 flags.go:64] FLAG: --image-service-endpoint="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082230 2171 flags.go:64] FLAG: --iptables-drop-bit="15" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082235 2171 flags.go:64] FLAG: --iptables-masquerade-bit="14" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082240 2171 flags.go:64] FLAG: --keep-terminated-pod-volumes="false" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082246 2171 flags.go:64] FLAG: --kernel-memcg-notification="false" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082250 2171 flags.go:64] FLAG: --kube-api-burst="10" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082255 2171 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082261 2171 flags.go:64] FLAG: --kube-api-qps="5" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082266 2171 flags.go:64] FLAG: --kube-reserved="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082279 2171 flags.go:64] FLAG: --kube-reserved-cgroup="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082285 2171 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082292 2171 flags.go:64] FLAG: --kubelet-cgroups="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082296 2171 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082302 2171 flags.go:64] FLAG: --lock-file="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082307 2171 flags.go:64] FLAG: --log-backtrace-at=":0" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082313 2171 flags.go:64] FLAG: --log-cadvisor-usage="false" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082319 2171 flags.go:64] FLAG: --log-dir="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082324 2171 flags.go:64] FLAG: --log-file="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082328 2171 flags.go:64] FLAG: --log-file-max-size="1800" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082334 2171 flags.go:64] FLAG: --log-flush-frequency="5s" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082340 2171 flags.go:64] FLAG: --log-json-info-buffer-size="0" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082348 2171 flags.go:64] FLAG: --log-json-split-stream="false" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082354 2171 flags.go:64] FLAG: --logging-format="text" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082359 2171 flags.go:64] FLAG: --logtostderr="true" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082376 2171 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082383 2171 flags.go:64] FLAG: --make-iptables-util-chains="true" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082388 2171 flags.go:64] FLAG: --manifest-url="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082393 2171 flags.go:64] FLAG: --manifest-url-header="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082400 2171 flags.go:64] FLAG: --master-service-namespace="default" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082409 2171 flags.go:64] FLAG: --max-housekeeping-interval="15s" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082416 2171 flags.go:64] FLAG: --max-open-files="1000000" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082422 2171 flags.go:64] FLAG: --max-pods="110" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082427 2171 flags.go:64] FLAG: --maximum-dead-containers="-1" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082433 2171 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082439 2171 flags.go:64] FLAG: --memory-manager-policy="None" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082444 2171 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082449 2171 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082454 2171 flags.go:64] FLAG: --node-ip="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082459 2171 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/worker=,node.openshift.io/os_id=rhcos" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082470 2171 flags.go:64] FLAG: --node-status-max-images="50" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082475 2171 flags.go:64] FLAG: --node-status-update-frequency="10s" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082482 2171 flags.go:64] FLAG: --one-output="false" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082488 2171 flags.go:64] FLAG: --oom-score-adj="-999" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082493 2171 flags.go:64] FLAG: --pod-cidr="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082498 2171 flags.go:64] FLAG: --pod-infra-container-image="registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082507 2171 flags.go:64] FLAG: --pod-manifest-path="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082512 2171 flags.go:64] FLAG: --pod-max-pids="-1" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082517 2171 flags.go:64] FLAG: --pods-per-core="0" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082527 2171 flags.go:64] FLAG: --port="10250" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082533 2171 flags.go:64] FLAG: --protect-kernel-defaults="false" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082538 2171 flags.go:64] FLAG: --provider-id="alicloud://us-east-1.i-0xif5dui7khr4rm1wp68" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082544 2171 flags.go:64] FLAG: --qos-reserved="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082550 2171 flags.go:64] FLAG: --read-only-port="10255" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082555 2171 flags.go:64] FLAG: --register-node="true" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082560 2171 flags.go:64] FLAG: --register-schedulable="true" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082566 2171 flags.go:64] FLAG: --register-with-taints="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082572 2171 flags.go:64] FLAG: --registry-burst="10" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082578 2171 flags.go:64] FLAG: --registry-qps="5" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082584 2171 flags.go:64] FLAG: --reserved-cpus="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082589 2171 flags.go:64] FLAG: --reserved-memory="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082596 2171 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082601 2171 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082607 2171 flags.go:64] FLAG: --rotate-certificates="false" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082612 2171 flags.go:64] FLAG: --rotate-server-certificates="false" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082617 2171 flags.go:64] FLAG: --runonce="false" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082622 2171 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082628 2171 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082633 2171 flags.go:64] FLAG: --seccomp-default="false" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082644 2171 flags.go:64] FLAG: --serialize-image-pulls="true" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082650 2171 flags.go:64] FLAG: --skip-headers="false" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082655 2171 flags.go:64] FLAG: --skip-log-headers="false" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082661 2171 flags.go:64] FLAG: --stderrthreshold="2" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082666 2171 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082672 2171 flags.go:64] FLAG: --storage-driver-db="cadvisor" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082678 2171 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082684 2171 flags.go:64] FLAG: --storage-driver-password="root" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082689 2171 flags.go:64] FLAG: --storage-driver-secure="false" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082694 2171 flags.go:64] FLAG: --storage-driver-table="stats" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082699 2171 flags.go:64] FLAG: --storage-driver-user="root" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082704 2171 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082709 2171 flags.go:64] FLAG: --sync-frequency="1m0s" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082715 2171 flags.go:64] FLAG: --system-cgroups="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082720 2171 flags.go:64] FLAG: --system-reserved="cpu=500m,ephemeral-storage=1Gi,memory=1Gi" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082729 2171 flags.go:64] FLAG: --system-reserved-cgroup="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082734 2171 flags.go:64] FLAG: --tls-cert-file="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082739 2171 flags.go:64] FLAG: --tls-cipher-suites="[]" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082752 2171 flags.go:64] FLAG: --tls-min-version="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082761 2171 flags.go:64] FLAG: --tls-private-key-file="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082768 2171 flags.go:64] FLAG: --topology-manager-policy="none" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082773 2171 flags.go:64] FLAG: --topology-manager-scope="container" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082779 2171 flags.go:64] FLAG: --v="2" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082786 2171 flags.go:64] FLAG: --version="false" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082794 2171 flags.go:64] FLAG: --vmodule="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082800 2171 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082806 2171 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.082928 2171 feature_gate.go:246] feature gates: &{map[APIPriorityAndFairness:true CSIMigrationAzureFile:false CSIMigrationvSphere:false DownwardAPIHugePages:true RotateKubeletServerCertificate:true]} Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.087491 2171 server.go:413] "Kubelet version" kubeletVersion="v1.25.4+77bec7a" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.087512 2171 server.go:415] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.087604 2171 feature_gate.go:246] feature gates: &{map[APIPriorityAndFairness:true CSIMigrationAzureFile:false CSIMigrationvSphere:false DownwardAPIHugePages:true RotateKubeletServerCertificate:true]} Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.087729 2171 feature_gate.go:246] feature gates: &{map[APIPriorityAndFairness:true CSIMigrationAzureFile:false CSIMigrationvSphere:false DownwardAPIHugePages:true RotateKubeletServerCertificate:true]} Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.087837 2171 server.go:825] "Client rotation is on, will bootstrap in background" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.096970 2171 bootstrap.go:84] "Current kubeconfig file contents are still valid, no bootstrap necessary" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.097658 2171 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.098041 2171 server.go:882] "Starting client certificate rotation" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.098060 2171 certificate_manager.go:270] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.100119 2171 certificate_manager.go:270] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2023-01-06 06:46:04 +0000 UTC, rotation deadline is 2023-01-06 03:18:40.255589238 +0000 UTC Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.100144 2171 certificate_manager.go:270] kubernetes.io/kube-apiserver-client-kubelet: Waiting 10h2m18.155447699s for next certificate rotation Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.106959 2171 dynamic_cafile_content.go:119] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.107157 2171 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.108807 2171 manager.go:163] cAdvisor running in container: "/system.slice/kubelet.service" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.114590 2171 fs.go:133] Filesystem UUIDs: map[249B-6B8C:/dev/vda2 7e283ec8-4fd4-48d9-bedf-150e5fbbb374:/dev/vda3 fda260a9-581c-4391-b124-5cfcba02ef65:/dev/vda4] Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.114607 2171 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /sys/fs/cgroup:{mountpoint:/sys/fs/cgroup major:0 minor:25 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:43 fsType:tmpfs blockSize:0}] Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.115194 2171 nvidia.go:54] NVIDIA GPU metrics disabled Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.116934 2171 manager.go:212] Machine: {Timestamp:2023-01-05 17:16:22.116782598 +0000 UTC m=+0.313027276 CPUVendorID:GenuineIntel NumCores:2 NumPhysicalCores:1 NumSockets:1 CpuFrequency:2500000 MemoryCapacity:8192086016 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:2bf656612c3249b68a06a3657cb09103 SystemUUID:2bf65661-2c32-49b6-8a06-a3657cb09103 BootID:dbcc4815-12e4-479e-bd59-e6b0facd2d6e Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:4096040960 Type:vfs Inodes:1000010 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:4096040960 Type:vfs Inodes:1000010 HasInodes:true} {Device:/sys/fs/cgroup DeviceMajor:0 DeviceMinor:25 Capacity:4096040960 Type:vfs Inodes:1000010 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:128300593152 Type:vfs Inodes:62651840 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:43 Capacity:4096040960 Type:vfs Inodes:1000010 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:128849018880 Scheduler:mq-deadline}] NetworkDevices:[{Name:br-ex MacAddress:00:16:3e:00:58:a6 Speed:0 Mtu:1500} {Name:ens5 MacAddress:00:16:3e:00:58:a6 Speed:-1 Mtu:1500} {Name:ovs-system MacAddress:8e:43:2e:2a:6c:c4 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:8192086016 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0 1] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0}] Caches:[{Id:0 Size:37486592 Type:Unified Level:3}]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.117090 2171 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.117287 2171 manager.go:228] Version: {KernelVersion:4.18.0-372.39.1.el8_6.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 412.86.202212170457-0 (Ootpa) DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.117777 2171 container_manager_linux.go:262] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.117860 2171 container_manager_linux.go:267] "Creating Container Manager object based on Node Config" nodeConfig={RuntimeCgroupsName:/system.slice/crio.service SystemCgroupsName:/system.slice KubeletCgroupsName: KubeletOOMScoreAdj:-999 ContainerRuntime: CgroupsPerQOS:true CgroupRoot:/ CgroupDriver:systemd KubeletRootDir:/var/lib/kubelet ProtectKernelDefaults:false NodeAllocatableConfig:{KubeReservedCgroupName: SystemReservedCgroupName: ReservedSystemCPUs: EnforceNodeAllocatable:map[pods:{}] KubeReserved:map[] SystemReserved:map[cpu:{i:{value:500 scale:-3} d:{Dec:} s:500m Format:DecimalSI} ephemeral-storage:{i:{value:1073741824 scale:0} d:{Dec:} s:1Gi Format:BinarySI} memory:{i:{value:1073741824 scale:0} d:{Dec:} s:1Gi Format:BinarySI}] HardEvictionThresholds:[{Signal:imagefs.available Operator:LessThan Value:{Quantity: Percentage:0.15} GracePeriod:0s MinReclaim:} {Signal:memory.available Operator:LessThan Value:{Quantity:100Mi Percentage:0} GracePeriod:0s MinReclaim:} {Signal:nodefs.available Operator:LessThan Value:{Quantity: Percentage:0.1} GracePeriod:0s MinReclaim:} {Signal:nodefs.inodesFree Operator:LessThan Value:{Quantity: Percentage:0.05} GracePeriod:0s MinReclaim:}]} QOSReserved:map[] ExperimentalCPUManagerPolicy:none ExperimentalCPUManagerPolicyOptions:map[] ExperimentalTopologyManagerScope:container ExperimentalCPUManagerReconcilePeriod:10s ExperimentalMemoryManagerPolicy:None ExperimentalMemoryManagerReservedMemory:[] ExperimentalPodPidsLimit:4096 EnforceCPULimits:true CPUCFSQuotaPeriod:100ms ExperimentalTopologyManagerPolicy:none} Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.117887 2171 topology_manager.go:134] "Creating topology manager with policy per scope" topologyPolicyName="none" topologyScopeName="container" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.117902 2171 container_manager_linux.go:302] "Creating device plugin manager" devicePluginEnabled=true Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.118535 2171 manager.go:127] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.118554 2171 server.go:64] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.118697 2171 state_mem.go:36] "Initialized new in-memory state store" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.118767 2171 util_unix.go:104] "Using this endpoint is deprecated, please consider using full URL format" endpoint="/var/run/crio/crio.sock" URL="unix:///var/run/crio/crio.sock" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.128796 2171 remote_runtime.go:139] "Using CRI v1 runtime API" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.128831 2171 util_unix.go:104] "Using this endpoint is deprecated, please consider using full URL format" endpoint="/var/run/crio/crio.sock" URL="unix:///var/run/crio/crio.sock" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.149039 2171 remote_image.go:95] "Using CRI v1 image API" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.149223 2171 server.go:1136] "Using root directory" path="/var/lib/kubelet" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.150724 2171 kubelet.go:393] "Attempting to sync node with API server" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.150756 2171 kubelet.go:282] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.150792 2171 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.150806 2171 kubelet.go:293] "Adding apiserver pod source" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.150835 2171 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.153605 2171 kuberuntime_manager.go:240] "Container runtime initialized" containerRuntime="cri-o" version="1.25.1-5.rhaos4.12.git6005903.el8" apiVersion="v1" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.154323 2171 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.156649 2171 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/vsphere-volume" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.156675 2171 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.156686 2171 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/rbd" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.156696 2171 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/aws-ebs" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.156707 2171 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/gce-pd" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.156718 2171 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/cinder" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.156729 2171 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/azure-disk" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.156741 2171 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/azure-file" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.157121 2171 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.157135 2171 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.157146 2171 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.157157 2171 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.157199 2171 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/secret" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.157217 2171 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.157229 2171 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/glusterfs" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.157244 2171 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/cephfs" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.157256 2171 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.157271 2171 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/fc" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.157283 2171 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.157295 2171 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/projected" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.157306 2171 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.157340 2171 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/csi" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.157538 2171 server.go:1175] "Started kubelet" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Started Kubernetes Kubelet. Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.159218 2171 certificate_manager.go:270] kubernetes.io/kubelet-serving: Certificate rotation is enabled Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.159239 2171 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.162397 2171 server.go:155] "Starting to listen" address="0.0.0.0" port=10250 Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.163555 2171 server.go:438] "Adding debug handlers to kubelet server" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: E0105 17:16:22.172080 2171 kubelet.go:1333] "Image garbage collection failed once. Stats initialization may not have completed yet" err="failed to get imageFs info: unable to find data in memory cache" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.173184 2171 certificate_manager.go:270] kubernetes.io/kubelet-serving: Certificate expiration is 2023-01-06 06:46:04 +0000 UTC, rotation deadline is 2023-01-06 02:38:05.736165117 +0000 UTC Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.173202 2171 certificate_manager.go:270] kubernetes.io/kubelet-serving: Waiting 9h21m43.562965546s for next certificate rotation Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:22.175467271Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=cc4105ad-4257-47f7-94a3-daf7e992863f name=/runtime.v1.ImageService/ImageStatus Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 17:16:22.180207329Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=cc4105ad-4257-47f7-94a3-daf7e992863f name=/runtime.v1.ImageService/ImageStatus Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.182277 2171 volume_manager.go:291] "The desired_state_of_world populator starts" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.182299 2171 volume_manager.go:293] "Starting Kubelet Volume Manager" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.184856 2171 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.204519 2171 factory.go:153] Registering CRI-O factory Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.204544 2171 factory.go:55] Registering systemd factory Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.204731 2171 factory.go:103] Registering Raw factory Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.204811 2171 manager.go:1201] Started watching for new ooms in manager Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.213834 2171 manager.go:302] Starting recovery of all containers Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.290337 2171 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.340071 2171 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.340584 2171 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.340610 2171 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.340646 2171 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.350840 2171 manager.go:307] Recovery completed Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.372307 2171 kubelet_node_status.go:110] "Node was previously registered" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.372637 2171 kubelet_node_status.go:75] "Successfully registered node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.401897 2171 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.402132 2171 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.402240 2171 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.402334 2171 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeNotReady" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.402407 2171 setters.go:545] "Node became not ready" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" condition={Type:Ready Status:False LastHeartbeatTime:2023-01-05 17:16:22.40231606 +0000 UTC m=+0.598560730 LastTransitionTime:2023-01-05 17:16:22.40231606 +0000 UTC m=+0.598560730 Reason:KubeletNotReady Message:[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]} Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.402426 2171 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeNotSchedulable" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.443279 2171 cpu_manager.go:214] "Starting CPU manager" policy="none" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.443302 2171 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.443319 2171 state_mem.go:36] "Initialized new in-memory state store" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.445430 2171 policy_none.go:49] "None policy: Start" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.446019 2171 memory_manager.go:168] "Starting memorymanager" policy="None" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.446126 2171 state_mem.go:35] "Initializing new in-memory state store" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.447604 2171 container_manager_linux.go:427] "Updating kernel flag" flag="vm/overcommit_memory" expectedValue=1 actualValue=0 Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.447853 2171 container_manager_linux.go:427] "Updating kernel flag" flag="kernel/panic" expectedValue=10 actualValue=0 Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.472411 2171 kubelet_network_linux.go:63] "Initialized iptables rules." protocol=IPv4 Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.494057 2171 manager.go:273] "Starting Device Plugin manager" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.494098 2171 manager.go:447] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.494110 2171 server.go:77] "Starting device plugin registration server" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.494292 2171 plugin_watcher.go:52] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.494358 2171 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.494384 2171 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.542974 2171 kubelet_network_linux.go:63] "Initialized iptables rules." protocol=IPv6 Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.543025 2171 status_manager.go:161] "Starting to sync pod status with apiserver" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.543042 2171 kubelet.go:2033] "Starting kubelet main sync loop" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: E0105 17:16:22.543132 2171 kubelet.go:2057] "Skipping pod synchronization" err="PLEG is not healthy: pleg has yet to be successful" Jan 05 17:16:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:22.644155 2171 kubelet.go:2119] "SyncLoop ADD" source="file" pods=[] Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.152771 2171 apiserver.go:52] "Watching apiserver" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.157387 2171 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-ingress-canary/ingress-canary-8xrbh openshift-dns/node-resolver-nxzr8 openshift-monitoring/node-exporter-2r6nf openshift-image-registry/node-ca-xw6d2 openshift-multus/multus-twl8f openshift-multus/cni-sysctl-allowlist-ds-tpxqs openshift-network-diagnostics/network-check-target-xmq2g openshift-sdn/sdn-fbccx openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4 openshift-cluster-node-tuning-operator/tuned-smwj7 openshift-dns/dns-default-2pfzf openshift-multus/network-metrics-daemon-hq594 openshift-machine-config-operator/machine-config-daemon-hf8f5 openshift-multus/multus-additional-cni-plugins-5fgxc] Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.157462 2171 topology_manager.go:205] "Topology Admit Handler" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.157553 2171 topology_manager.go:205] "Topology Admit Handler" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.157624 2171 topology_manager.go:205] "Topology Admit Handler" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.157693 2171 topology_manager.go:205] "Topology Admit Handler" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.157750 2171 topology_manager.go:205] "Topology Admit Handler" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.157803 2171 topology_manager.go:205] "Topology Admit Handler" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.157891 2171 topology_manager.go:205] "Topology Admit Handler" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.159130 2171 topology_manager.go:205] "Topology Admit Handler" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.159314 2171 topology_manager.go:205] "Topology Admit Handler" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.159501 2171 topology_manager.go:205] "Topology Admit Handler" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.159644 2171 topology_manager.go:205] "Topology Admit Handler" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.159805 2171 topology_manager.go:205] "Topology Admit Handler" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.159945 2171 topology_manager.go:205] "Topology Admit Handler" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.160108 2171 topology_manager.go:205] "Topology Admit Handler" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.204101 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cookie-secret\" (UniqueName: \"kubernetes.io/secret/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-cookie-secret\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.204148 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-cnibin\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.204184 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8f88cbe5-1221-4e6f-a6c9-f22da152b43f-hosts-file\") pod \"node-resolver-nxzr8\" (UID: \"8f88cbe5-1221-4e6f-a6c9-f22da152b43f\") " pod="openshift-dns/node-resolver-nxzr8" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.204220 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-lib-modules\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.204270 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-wtmp\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.204307 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-os-release\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.204356 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wq4k7\" (UniqueName: \"kubernetes.io/projected/c0e47249-a5ff-460e-958d-53577387c154-kube-api-access-wq4k7\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.204417 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.204556 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-etc\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.204601 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfs2p\" (UniqueName: \"kubernetes.io/projected/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-kube-api-access-dfs2p\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.204660 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7www\" (UniqueName: \"kubernetes.io/projected/8f88cbe5-1221-4e6f-a6c9-f22da152b43f-kube-api-access-g7www\") pod \"node-resolver-nxzr8\" (UID: \"8f88cbe5-1221-4e6f-a6c9-f22da152b43f\") " pod="openshift-dns/node-resolver-nxzr8" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.204704 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-mtu-migration\" (UniqueName: \"kubernetes.io/configmap/b5d621a5-07d9-459b-8745-a991cf4a179e-config-mtu-migration\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.204741 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-etc-sysconfig\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.204784 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xrks\" (UniqueName: \"kubernetes.io/projected/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-kube-api-access-5xrks\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.204827 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-metrics-client-ca\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.204867 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-run-crio\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-crio\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.204918 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-run-netns\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.204968 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-run-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-openshift-sdn\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.205032 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-plugin-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.205075 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx2gt\" (UniqueName: \"kubernetes.io/projected/50152296-9489-4fa8-aa42-7858debe1c08-kube-api-access-zx2gt\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.205110 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-modules\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-modules\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.205159 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sdn-metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5d621a5-07d9-459b-8745-a991cf4a179e-sdn-metrics-certs\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.205210 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-registration-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.205262 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-proxy-tls\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.205305 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd-system\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-run-systemd-system\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.205343 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-tuned-profiles-data\" (UniqueName: \"kubernetes.io/configmap/91265f01-9dcb-4046-961e-ad4c544224d7-var-lib-tuned-profiles-data\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.205397 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-os-release\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.205438 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-dbus\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.205510 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-cni-bin\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.205553 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g48vj\" (UniqueName: \"kubernetes.io/projected/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-kube-api-access-g48vj\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.205598 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-rootfs\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.205635 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-host\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.205691 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmxtc\" (UniqueName: \"kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc\") pod \"network-check-target-xmq2g\" (UID: \"f364a949-09d7-4cfb-83ff-e532b822a557\") " pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.205729 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-slash\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.205773 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-metrics-tls\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.205811 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-multus-cni-dir\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.205866 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.205912 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzfn5\" (UniqueName: \"kubernetes.io/projected/e04a47b4-aa8e-4d4e-9d45-aa8037dcb748-kube-api-access-lzfn5\") pod \"ingress-canary-8xrbh\" (UID: \"e04a47b4-aa8e-4d4e-9d45-aa8037dcb748\") " pod="openshift-ingress-canary/ingress-canary-8xrbh" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.205953 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-system-cni-dir\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.206015 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-run-ovs\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-ovs\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.206129 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7xz4\" (UniqueName: \"kubernetes.io/projected/91265f01-9dcb-4046-961e-ad4c544224d7-kube-api-access-b7xz4\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.206170 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-kubelet-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.206208 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-cnibin\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.206248 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/659b8bb4-81b5-49bf-8127-1d326f30ba2f-host\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.206291 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"non-standard-root-system-trust-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bb84a70-232c-4467-ac0b-647030155908-non-standard-root-system-trust-ca-bundle\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.206330 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-sys\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.206390 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-sys\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.206438 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e13f5318-440f-4b4e-863d-2489155b9feb-cni-binary-copy\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.206477 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/659b8bb4-81b5-49bf-8127-1d326f30ba2f-serviceca\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.206520 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4sjn\" (UniqueName: \"kubernetes.io/projected/659b8bb4-81b5-49bf-8127-1d326f30ba2f-kube-api-access-h4sjn\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.206558 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b5d621a5-07d9-459b-8745-a991cf4a179e-env-overrides\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.206647 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-config-volume\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.206686 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-tls\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.206726 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-system-cni-dir\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.206769 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9799x\" (UniqueName: \"kubernetes.io/projected/e13f5318-440f-4b4e-863d-2489155b9feb-kube-api-access-9799x\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.206810 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c0e47249-a5ff-460e-958d-53577387c154-cni-binary-copy\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.206851 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7d6lx\" (UniqueName: \"kubernetes.io/projected/b5d621a5-07d9-459b-8745-a991cf4a179e-kube-api-access-7d6lx\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.206891 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-var-run-dbus\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.206931 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-root\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.206974 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-textfile\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.207036 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.207081 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c0e47249-a5ff-460e-958d-53577387c154-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.207128 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5d621a5-07d9-459b-8745-a991cf4a179e-config\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.207231 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-device-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.207287 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-run-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-kubernetes\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.207327 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-conf\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-cni-conf\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.207383 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-lib-cni-networks-openshift-sdn\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.207438 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzzxr\" (UniqueName: \"kubernetes.io/projected/1bb84a70-232c-4467-ac0b-647030155908-kube-api-access-vzzxr\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.207451 2171 reconciler.go:169] "Reconciler: start to sync state" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.211574 2171 reconciler.go:537] "Reconciler sync states: could not find pod information in desired state, update it in actual state" reconstructedVolume=&{volumeName:kubernetes.io/configmap/02e0075b-c660-4dc5-8cce-00fc4ef2980f-cni-sysctl-allowlist podName:02e0075b-c660-4dc5-8cce-00fc4ef2980f volumeSpec:0xc00107d3c8 outerVolumeSpecName:cni-sysctl-allowlist pod:0xc000b09800 volumeGidValue: devicePath: mounter:0xc0010f5b00 deviceMounter: blockVolumeMapper:} Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.211641 2171 reconciler.go:537] "Reconciler sync states: could not find pod information in desired state, update it in actual state" reconstructedVolume=&{volumeName:kubernetes.io/empty-dir/02e0075b-c660-4dc5-8cce-00fc4ef2980f-ready podName:02e0075b-c660-4dc5-8cce-00fc4ef2980f volumeSpec:0xc00107d3f8 outerVolumeSpecName:ready pod:0xc000b09c00 volumeGidValue: devicePath: mounter:0xc0008cb490 deviceMounter: blockVolumeMapper:} Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.211706 2171 reconciler.go:537] "Reconciler sync states: could not find pod information in desired state, update it in actual state" reconstructedVolume=&{volumeName:kubernetes.io/projected/02e0075b-c660-4dc5-8cce-00fc4ef2980f-kube-api-access-7fwhf podName:02e0075b-c660-4dc5-8cce-00fc4ef2980f volumeSpec:0xc00107d410 outerVolumeSpecName:kube-api-access-7fwhf pod:0xc001a0a000 volumeGidValue: devicePath: mounter:0xc0019f7980 deviceMounter: blockVolumeMapper:} Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.318900 2171 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fwhf\" (UniqueName: \"kubernetes.io/projected/02e0075b-c660-4dc5-8cce-00fc4ef2980f-kube-api-access-7fwhf\") pod \"02e0075b-c660-4dc5-8cce-00fc4ef2980f\" (UID: \"02e0075b-c660-4dc5-8cce-00fc4ef2980f\") " Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.318934 2171 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/02e0075b-c660-4dc5-8cce-00fc4ef2980f-cni-sysctl-allowlist\") pod \"02e0075b-c660-4dc5-8cce-00fc4ef2980f\" (UID: \"02e0075b-c660-4dc5-8cce-00fc4ef2980f\") " Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.318968 2171 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/02e0075b-c660-4dc5-8cce-00fc4ef2980f-ready\") pod \"02e0075b-c660-4dc5-8cce-00fc4ef2980f\" (UID: \"02e0075b-c660-4dc5-8cce-00fc4ef2980f\") " Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.319063 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-cnibin\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.319100 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/659b8bb4-81b5-49bf-8127-1d326f30ba2f-host\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.319130 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b5d621a5-07d9-459b-8745-a991cf4a179e-env-overrides\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.319170 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-config-volume\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.319205 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"non-standard-root-system-trust-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bb84a70-232c-4467-ac0b-647030155908-non-standard-root-system-trust-ca-bundle\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.319239 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-sys\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.319268 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-sys\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.319297 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e13f5318-440f-4b4e-863d-2489155b9feb-cni-binary-copy\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.319327 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/659b8bb4-81b5-49bf-8127-1d326f30ba2f-serviceca\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.319357 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-h4sjn\" (UniqueName: \"kubernetes.io/projected/659b8bb4-81b5-49bf-8127-1d326f30ba2f-kube-api-access-h4sjn\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.319431 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-tls\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.319462 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-system-cni-dir\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.319492 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-9799x\" (UniqueName: \"kubernetes.io/projected/e13f5318-440f-4b4e-863d-2489155b9feb-kube-api-access-9799x\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.319525 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c0e47249-a5ff-460e-958d-53577387c154-cni-binary-copy\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.319568 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-7d6lx\" (UniqueName: \"kubernetes.io/projected/b5d621a5-07d9-459b-8745-a991cf4a179e-kube-api-access-7d6lx\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.319600 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5d621a5-07d9-459b-8745-a991cf4a179e-config\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.319630 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-device-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.319665 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-var-run-dbus\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.319716 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-root\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.319757 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-textfile\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.319795 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.319833 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c0e47249-a5ff-460e-958d-53577387c154-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.319864 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-run-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-kubernetes\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.319893 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-cni-conf\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-cni-conf\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.319927 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-lib-cni-networks-openshift-sdn\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.319961 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-vzzxr\" (UniqueName: \"kubernetes.io/projected/1bb84a70-232c-4467-ac0b-647030155908-kube-api-access-vzzxr\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.320024 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cookie-secret\" (UniqueName: \"kubernetes.io/secret/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-cookie-secret\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.320059 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-cnibin\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.320097 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8f88cbe5-1221-4e6f-a6c9-f22da152b43f-hosts-file\") pod \"node-resolver-nxzr8\" (UID: \"8f88cbe5-1221-4e6f-a6c9-f22da152b43f\") " pod="openshift-dns/node-resolver-nxzr8" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.320126 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-lib-modules\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.320165 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-wtmp\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.320199 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-os-release\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.320233 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-wq4k7\" (UniqueName: \"kubernetes.io/projected/c0e47249-a5ff-460e-958d-53577387c154-kube-api-access-wq4k7\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.320265 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.320293 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"etc\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-etc\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.320334 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-dfs2p\" (UniqueName: \"kubernetes.io/projected/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-kube-api-access-dfs2p\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.320388 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-g7www\" (UniqueName: \"kubernetes.io/projected/8f88cbe5-1221-4e6f-a6c9-f22da152b43f-kube-api-access-g7www\") pod \"node-resolver-nxzr8\" (UID: \"8f88cbe5-1221-4e6f-a6c9-f22da152b43f\") " pod="openshift-dns/node-resolver-nxzr8" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.320426 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config-mtu-migration\" (UniqueName: \"kubernetes.io/configmap/b5d621a5-07d9-459b-8745-a991cf4a179e-config-mtu-migration\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.320470 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-etc-sysconfig\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.320516 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"plugin-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-plugin-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.320553 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-5xrks\" (UniqueName: \"kubernetes.io/projected/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-kube-api-access-5xrks\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.320584 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-metrics-client-ca\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.320614 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-run-crio\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-crio\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.320649 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-run-netns\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.320679 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-run-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-openshift-sdn\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.320719 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-zx2gt\" (UniqueName: \"kubernetes.io/projected/50152296-9489-4fa8-aa42-7858debe1c08-kube-api-access-zx2gt\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.320749 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-modules\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-modules\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.320779 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"sdn-metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5d621a5-07d9-459b-8745-a991cf4a179e-sdn-metrics-certs\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.320811 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-cni-bin\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.320848 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-g48vj\" (UniqueName: \"kubernetes.io/projected/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-kube-api-access-g48vj\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.320888 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-registration-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.320923 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-proxy-tls\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.320959 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"run-systemd-system\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-run-systemd-system\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.321016 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"var-lib-tuned-profiles-data\" (UniqueName: \"kubernetes.io/configmap/91265f01-9dcb-4046-961e-ad4c544224d7-var-lib-tuned-profiles-data\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.321055 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-os-release\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.321086 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-dbus\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.321122 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-rootfs\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.321155 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-host\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.321195 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-mmxtc\" (UniqueName: \"kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc\") pod \"network-check-target-xmq2g\" (UID: \"f364a949-09d7-4cfb-83ff-e532b822a557\") " pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.321228 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-slash\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.321258 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-metrics-tls\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.321287 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-multus-cni-dir\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.321324 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.321379 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-lzfn5\" (UniqueName: \"kubernetes.io/projected/e04a47b4-aa8e-4d4e-9d45-aa8037dcb748-kube-api-access-lzfn5\") pod \"ingress-canary-8xrbh\" (UID: \"e04a47b4-aa8e-4d4e-9d45-aa8037dcb748\") " pod="openshift-ingress-canary/ingress-canary-8xrbh" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.321416 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-system-cni-dir\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.321470 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-run-ovs\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-ovs\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.321511 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-b7xz4\" (UniqueName: \"kubernetes.io/projected/91265f01-9dcb-4046-961e-ad4c544224d7-kube-api-access-b7xz4\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.321550 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-kubelet-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.321756 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-kubelet-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 17:16:23.323888 2171 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/02e0075b-c660-4dc5-8cce-00fc4ef2980f/volumes/kubernetes.io~projected/kube-api-access-7fwhf: clearQuota called, but quotas disabled Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.323978 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-system-cni-dir\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.324211 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-cnibin\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.324262 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/659b8bb4-81b5-49bf-8127-1d326f30ba2f-host\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.324490 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b5d621a5-07d9-459b-8745-a991cf4a179e-env-overrides\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 17:16:23.324742 2171 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/02e0075b-c660-4dc5-8cce-00fc4ef2980f/volumes/kubernetes.io~configmap/cni-sysctl-allowlist: clearQuota called, but quotas disabled Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.325292 2171 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02e0075b-c660-4dc5-8cce-00fc4ef2980f-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "02e0075b-c660-4dc5-8cce-00fc4ef2980f" (UID: "02e0075b-c660-4dc5-8cce-00fc4ef2980f"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.325934 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5d621a5-07d9-459b-8745-a991cf4a179e-config\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.326056 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-device-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.326395 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"root\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-root\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.326512 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-textfile\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.327218 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-sys\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.327498 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-sys\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.327868 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"non-standard-root-system-trust-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bb84a70-232c-4467-ac0b-647030155908-non-standard-root-system-trust-ca-bundle\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.328463 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c0e47249-a5ff-460e-958d-53577387c154-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.329129 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-run-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-kubernetes\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.329203 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-cni-conf\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-cni-conf\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 17:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.329259 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-lib-cni-networks-openshift-sdn\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.330167 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-cnibin\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.330254 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8f88cbe5-1221-4e6f-a6c9-f22da152b43f-hosts-file\") pod \"node-resolver-nxzr8\" (UID: \"8f88cbe5-1221-4e6f-a6c9-f22da152b43f\") " pod="openshift-dns/node-resolver-nxzr8" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.330353 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-lib-modules\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.330481 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-wtmp\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 17:16:23.330567 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-os-release\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.684477 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"etc\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-etc\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.685181 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config-mtu-migration\" (UniqueName: \"kubernetes.io/configmap/b5d621a5-07d9-459b-8745-a991cf4a179e-config-mtu-migration\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.686220 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-etc-sysconfig\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.686308 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"plugin-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-plugin-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.687111 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-metrics-client-ca\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.687203 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-run-crio\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-crio\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.687260 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-run-netns\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.687330 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-run-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-openshift-sdn\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.687632 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-modules\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-modules\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.689802 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cookie-secret\" (UniqueName: \"kubernetes.io/secret/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-cookie-secret\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.689902 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e13f5318-440f-4b4e-863d-2489155b9feb-cni-binary-copy\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.690192 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/659b8bb4-81b5-49bf-8127-1d326f30ba2f-serviceca\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.694858 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"sdn-metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5d621a5-07d9-459b-8745-a991cf4a179e-sdn-metrics-certs\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.694935 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-cni-bin\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.695742 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-registration-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.700941 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-tls\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 09:16:23.701008 2171 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/02e0075b-c660-4dc5-8cce-00fc4ef2980f/volumes/kubernetes.io~empty-dir/ready: clearQuota called, but quotas disabled Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.706539 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-proxy-tls\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.706652 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"run-systemd-system\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-run-systemd-system\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.708369 2171 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02e0075b-c660-4dc5-8cce-00fc4ef2980f-kube-api-access-7fwhf" (OuterVolumeSpecName: "kube-api-access-7fwhf") pod "02e0075b-c660-4dc5-8cce-00fc4ef2980f" (UID: "02e0075b-c660-4dc5-8cce-00fc4ef2980f"). InnerVolumeSpecName "kube-api-access-7fwhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.708682 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-os-release\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.708799 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-dbus\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.708884 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-rootfs\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.708957 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-host\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.709269 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-slash\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.711417 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-var-run-dbus\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.711461 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"var-lib-tuned-profiles-data\" (UniqueName: \"kubernetes.io/configmap/91265f01-9dcb-4046-961e-ad4c544224d7-var-lib-tuned-profiles-data\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.711543 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-multus-cni-dir\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.714260 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.716895 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-system-cni-dir\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.718185 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-run-ovs\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-ovs\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.718489 2171 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02e0075b-c660-4dc5-8cce-00fc4ef2980f-ready" (OuterVolumeSpecName: "ready") pod "02e0075b-c660-4dc5-8cce-00fc4ef2980f" (UID: "02e0075b-c660-4dc5-8cce-00fc4ef2980f"). InnerVolumeSpecName "ready". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.718537 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-config-volume\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.718673 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.718718 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c0e47249-a5ff-460e-958d-53577387c154-cni-binary-copy\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.721086 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4sjn\" (UniqueName: \"kubernetes.io/projected/659b8bb4-81b5-49bf-8127-1d326f30ba2f-kube-api-access-h4sjn\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.726441 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-9799x\" (UniqueName: \"kubernetes.io/projected/e13f5318-440f-4b4e-863d-2489155b9feb-kube-api-access-9799x\") pod \"multus-twl8f\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " pod="openshift-multus/multus-twl8f" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.727237 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.730925 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-metrics-tls\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.739255 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-wq4k7\" (UniqueName: \"kubernetes.io/projected/c0e47249-a5ff-460e-958d-53577387c154-kube-api-access-wq4k7\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.741787 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7www\" (UniqueName: \"kubernetes.io/projected/8f88cbe5-1221-4e6f-a6c9-f22da152b43f-kube-api-access-g7www\") pod \"node-resolver-nxzr8\" (UID: \"8f88cbe5-1221-4e6f-a6c9-f22da152b43f\") " pod="openshift-dns/node-resolver-nxzr8" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.744867 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-7d6lx\" (UniqueName: \"kubernetes.io/projected/b5d621a5-07d9-459b-8745-a991cf4a179e-kube-api-access-7d6lx\") pod \"sdn-fbccx\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " pod="openshift-sdn/sdn-fbccx" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.750690 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzzxr\" (UniqueName: \"kubernetes.io/projected/1bb84a70-232c-4467-ac0b-647030155908-kube-api-access-vzzxr\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.754408 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx2gt\" (UniqueName: \"kubernetes.io/projected/50152296-9489-4fa8-aa42-7858debe1c08-kube-api-access-zx2gt\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.755421 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xrks\" (UniqueName: \"kubernetes.io/projected/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-kube-api-access-5xrks\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.759204 2171 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.760311840Z" level=info msg="Running pod sandbox: openshift-multus/network-metrics-daemon-hq594/POD" id=8baf2240-92cb-468d-a2fc-59548f7ca737 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.760591746Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.765127 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-g48vj\" (UniqueName: \"kubernetes.io/projected/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-kube-api-access-g48vj\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.773110 2171 reconciler.go:399] "Volume detached for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/02e0075b-c660-4dc5-8cce-00fc4ef2980f-ready\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.773138 2171 reconciler.go:399] "Volume detached for volume \"kube-api-access-7fwhf\" (UniqueName: \"kubernetes.io/projected/02e0075b-c660-4dc5-8cce-00fc4ef2980f-kube-api-access-7fwhf\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.774314 2171 reconciler.go:399] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/02e0075b-c660-4dc5-8cce-00fc4ef2980f-cni-sysctl-allowlist\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.777641 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfs2p\" (UniqueName: \"kubernetes.io/projected/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-kube-api-access-dfs2p\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.785512 2171 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-twl8f" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.787096714Z" level=info msg="Running pod sandbox: openshift-multus/multus-twl8f/POD" id=6dc846c7-3ada-4c5f-ae4a-c8c5478c5997 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.787178108Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.792635 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzfn5\" (UniqueName: \"kubernetes.io/projected/e04a47b4-aa8e-4d4e-9d45-aa8037dcb748-kube-api-access-lzfn5\") pod \"ingress-canary-8xrbh\" (UID: \"e04a47b4-aa8e-4d4e-9d45-aa8037dcb748\") " pod="openshift-ingress-canary/ingress-canary-8xrbh" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.793136 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7xz4\" (UniqueName: \"kubernetes.io/projected/91265f01-9dcb-4046-961e-ad4c544224d7-kube-api-access-b7xz4\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.799582 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmxtc\" (UniqueName: \"kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc\") pod \"network-check-target-xmq2g\" (UID: \"f364a949-09d7-4cfb-83ff-e532b822a557\") " pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.804580803Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=6dc846c7-3ada-4c5f-ae4a-c8c5478c5997 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.806664641Z" level=info msg="Got pod network &{Name:network-metrics-daemon-hq594 Namespace:openshift-multus ID:8e2d55780ef7738a05fee323e5c71bd87390c87bd4b55a419d3d7362eb940075 UID:50152296-9489-4fa8-aa42-7858debe1c08 NetNS:/var/run/netns/74a8e5e6-0f86-443d-a641-ea34c5ca1020 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.806698840Z" level=info msg="Adding pod openshift-multus_network-metrics-daemon-hq594 to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.813861 2171 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.815093070Z" level=info msg="Running pod sandbox: openshift-monitoring/node-exporter-2r6nf/POD" id=d0d0c811-78df-49d4-a151-e5fbc5004554 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.815745490Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 09:16:23.828734 2171 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode13f5318_440f_4b4e_863d_2489155b9feb.slice/crio-d04b2e4ec3af58e65ad07d97f2565affd7e50b26af21f25245500b2278268045.scope WatchSource:0}: Error finding container d04b2e4ec3af58e65ad07d97f2565affd7e50b26af21f25245500b2278268045: Status 404 returned error can't find the container with id d04b2e4ec3af58e65ad07d97f2565affd7e50b26af21f25245500b2278268045 Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: E0105 09:16:23.830636 2171 watcher.go:152] Failed to watch directory "/sys/fs/cgroup/devices/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode04a47b4_aa8e_4d4e_9d45_aa8037dcb748.slice": readdirent /sys/fs/cgroup/devices/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode04a47b4_aa8e_4d4e_9d45_aa8037dcb748.slice: no such file or directory Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.837987932Z" level=info msg="Ran pod sandbox d04b2e4ec3af58e65ad07d97f2565affd7e50b26af21f25245500b2278268045 with infra container: openshift-multus/multus-twl8f/POD" id=6dc846c7-3ada-4c5f-ae4a-c8c5478c5997 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.842828217Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=d0d0c811-78df-49d4-a151-e5fbc5004554 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.845306 2171 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.845657431Z" level=info msg="Running pod sandbox: openshift-network-diagnostics/network-check-target-xmq2g/POD" id=05d8b8df-127d-4078-a91a-6443ff5ac6fb name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.845790986Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.845904 2171 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-8xrbh" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.846482606Z" level=info msg="Running pod sandbox: openshift-ingress-canary/ingress-canary-8xrbh/POD" id=5b3b3170-533a-4d99-9f5e-955d1e695c6f name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.846615368Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.854814 2171 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-sdn/sdn-fbccx" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.855548983Z" level=info msg="Ran pod sandbox a87b4ab12cf54ed2e65342c3d9c4ea51c24387aaa25fa0e287ee0befc098a632 with infra container: openshift-monitoring/node-exporter-2r6nf/POD" id=d0d0c811-78df-49d4-a151-e5fbc5004554 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 09:16:23.856689 2171 watcher.go:93] Error while processing event ("/sys/fs/cgroup/devices/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5d621a5_07d9_459b_8745_a991cf4a179e.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/devices/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5d621a5_07d9_459b_8745_a991cf4a179e.slice: no such file or directory Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.866223246Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=57fc8046-5e6e-4458-a55d-10cb7595f829 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.866563433Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=57fc8046-5e6e-4458-a55d-10cb7595f829 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.867616978Z" level=info msg="Running pod sandbox: openshift-sdn/sdn-fbccx/POD" id=3edc0770-ab05-45b4-b118-ce0020572f58 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.867726342Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.874681683Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180" id=1242c8bd-5eaa-4122-82c5-da5c0fe680cb name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.874845330Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=160bf407-a17b-49e5-8e51-1b66c5c7779e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.877222282Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=160bf407-a17b-49e5-8e51-1b66c5c7779e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.877581767Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b4cde205857c1e0a984b38c61275a8370136432c7c84dbfbd7f4835f26bbb935,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180],Size_:297268865,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=1242c8bd-5eaa-4122-82c5-da5c0fe680cb name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.878613 2171 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.879673960Z" level=info msg="Running pod sandbox: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/POD" id=c280110a-f05a-4511-b49a-06a2e85c44aa name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.879809337Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.883796465Z" level=info msg="Creating container: openshift-multus/multus-twl8f/kube-multus" id=69821c55-1902-4c8a-9586-ea3ed4bb6fb8 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.883875043Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.885089985Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180" id=9b24b9e6-2795-4795-a8cb-9859cf88ae67 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.889151851Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b4cde205857c1e0a984b38c61275a8370136432c7c84dbfbd7f4835f26bbb935,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180],Size_:297268865,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=9b24b9e6-2795-4795-a8cb-9859cf88ae67 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.890249203Z" level=info msg="Creating container: openshift-monitoring/node-exporter-2r6nf/init-textfile" id=fa4c6480-e488-4423-bc3c-8ce3d0af81f5 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.890405602Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.891998 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-2r6nf" event=&{ID:aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f Type:ContainerStarted Data:a87b4ab12cf54ed2e65342c3d9c4ea51c24387aaa25fa0e287ee0befc098a632} Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.893370 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-twl8f" event=&{ID:e13f5318-440f-4b4e-863d-2489155b9feb Type:ContainerStarted Data:d04b2e4ec3af58e65ad07d97f2565affd7e50b26af21f25245500b2278268045} Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.900105083Z" level=info msg="Got pod network &{Name:network-check-target-xmq2g Namespace:openshift-network-diagnostics ID:00fb6557a9939b9a2bc7c25ec8cc1770b954010cd506656807f39f57849767a8 UID:f364a949-09d7-4cfb-83ff-e532b822a557 NetNS:/var/run/netns/4e35a0fe-47dd-4c20-9ad8-c4b066c02651 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.900132605Z" level=info msg="Adding pod openshift-network-diagnostics_network-check-target-xmq2g to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.902322 2171 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-xw6d2" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.902672173Z" level=info msg="Running pod sandbox: openshift-image-registry/node-ca-xw6d2/POD" id=d0b42c08-b1f8-40b0-9940-fb750d7163eb name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.902877706Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.911359325Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=3edc0770-ab05-45b4-b118-ce0020572f58 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.919804 2171 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.920215254Z" level=info msg="Running pod sandbox: openshift-multus/multus-additional-cni-plugins-5fgxc/POD" id=44eeae78-ad0e-4247-9002-93fd13f4f2d3 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.920314756Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.934882116Z" level=info msg="Ran pod sandbox 66a629012ef7e22f6af44b09d2cda79cab33d282fd659514ee53da6e316c1302 with infra container: openshift-sdn/sdn-fbccx/POD" id=3edc0770-ab05-45b4-b118-ce0020572f58 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.935889409Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:30387b6f2e605bb8d018e0c09f9d0275048fc5e7966137fbe6bd4307c88d3fdd" id=2de210d3-bec1-4f0d-ba0a-d1bb0279e56d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.936362631Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:cff39d57d2c205a584f8bf4299b1e125fa1a96ac92f680117ba82c20334754d1,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:30387b6f2e605bb8d018e0c09f9d0275048fc5e7966137fbe6bd4307c88d3fdd],Size_:701241664,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=2de210d3-bec1-4f0d-ba0a-d1bb0279e56d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.938814341Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:30387b6f2e605bb8d018e0c09f9d0275048fc5e7966137fbe6bd4307c88d3fdd" id=803d0ea2-b751-47de-8597-aa61afe6db10 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.939072760Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:cff39d57d2c205a584f8bf4299b1e125fa1a96ac92f680117ba82c20334754d1,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:30387b6f2e605bb8d018e0c09f9d0275048fc5e7966137fbe6bd4307c88d3fdd],Size_:701241664,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=803d0ea2-b751-47de-8597-aa61afe6db10 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.940854 2171 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-nxzr8" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.941600749Z" level=info msg="Running pod sandbox: openshift-dns/node-resolver-nxzr8/POD" id=e0cf5218-e82a-48b9-b699-96f3ccb269d4 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.941710289Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.941940576Z" level=info msg="Creating container: openshift-sdn/sdn-fbccx/sdn" id=77807cab-bd17-4e36-b2b9-daf0e3f17281 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.942060949Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.948674778Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=d0b42c08-b1f8-40b0-9940-fb750d7163eb name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.955468967Z" level=info msg="Got pod network &{Name:ingress-canary-8xrbh Namespace:openshift-ingress-canary ID:16c0a5a6c5f7e9eecc5461c963ba5b26d7e13b2c8571603f6e05ef86ce3505e3 UID:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748 NetNS:/var/run/netns/185ab30d-a13f-46e2-a94f-58153e8799e2 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.955494672Z" level=info msg="Adding pod openshift-ingress-canary_ingress-canary-8xrbh to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.955743 2171 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.960193 2171 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-multus/cni-sysctl-allowlist-ds-tpxqs] Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:23.965290 2171 kubelet.go:2129] "SyncLoop REMOVE" source="api" pods=[openshift-multus/cni-sysctl-allowlist-ds-tpxqs] Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.969255774Z" level=info msg="Running pod sandbox: openshift-machine-config-operator/machine-config-daemon-hf8f5/POD" id=317d0bb5-e742-4c97-be7b-794a1c4f7251 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.969405682Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 09:16:23.970300 2171 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod659b8bb4_81b5_49bf_8127_1d326f30ba2f.slice/crio-080616c6ee5360407adb7a3e03767428f4727a0ee1746976a16081b2970f8d20.scope WatchSource:0}: Error finding container 080616c6ee5360407adb7a3e03767428f4727a0ee1746976a16081b2970f8d20: Status 404 returned error can't find the container with id 080616c6ee5360407adb7a3e03767428f4727a0ee1746976a16081b2970f8d20 Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.976744836Z" level=info msg="Ran pod sandbox 080616c6ee5360407adb7a3e03767428f4727a0ee1746976a16081b2970f8d20 with infra container: openshift-image-registry/node-ca-xw6d2/POD" id=d0b42c08-b1f8-40b0-9940-fb750d7163eb name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.979052271Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8" id=50340956-a791-4493-832c-7966ba42486e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.983543396Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:c9e2954d8b9d9e1594d0a16b90086a9297ffd65773a6aa4a64f325e871ce83d1,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8],Size_:363009057,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=50340956-a791-4493-832c-7966ba42486e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.987286810Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8" id=2837862d-3375-4983-886a-8e1b2082cf8d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.987566040Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:c9e2954d8b9d9e1594d0a16b90086a9297ffd65773a6aa4a64f325e871ce83d1,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8],Size_:363009057,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=2837862d-3375-4983-886a-8e1b2082cf8d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.988421447Z" level=info msg="Creating container: openshift-image-registry/node-ca-xw6d2/node-ca" id=616715eb-122a-4891-90c8-a9935a97c31e name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:23.988491610Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.004040217Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=317d0bb5-e742-4c97-be7b-794a1c4f7251 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:24.030000 2171 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.030367250Z" level=info msg="Running pod sandbox: openshift-cluster-node-tuning-operator/tuned-smwj7/POD" id=8fac1130-098d-4317-b72e-20f50c13091c name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.030429523Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.042558361Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=c280110a-f05a-4511-b49a-06a2e85c44aa name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 09:16:24.045328 2171 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod997b7ab4_939e_465c_9c7d_4a2ebe3a797a.slice/crio-c41ae071597e07deb9c5ea067a36e06b0f2b19fc9c9d6633270660715cffd374.scope WatchSource:0}: Error finding container c41ae071597e07deb9c5ea067a36e06b0f2b19fc9c9d6633270660715cffd374: Status 404 returned error can't find the container with id c41ae071597e07deb9c5ea067a36e06b0f2b19fc9c9d6633270660715cffd374 Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:24.046467 2171 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-2pfzf" Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.047667375Z" level=info msg="Running pod sandbox: openshift-dns/dns-default-2pfzf/POD" id=4a35ce7e-1e21-4eef-8938-665128689f0f name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.047780275Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.052949241Z" level=info msg="Ran pod sandbox c41ae071597e07deb9c5ea067a36e06b0f2b19fc9c9d6633270660715cffd374 with infra container: openshift-machine-config-operator/machine-config-daemon-hf8f5/POD" id=317d0bb5-e742-4c97-be7b-794a1c4f7251 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.056043521Z" level=info msg="Ran pod sandbox 821f01442b5e38ec5d5958577a6f01e03591ea0d8fc60b05f51cb69558bd0cb1 with infra container: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/POD" id=c280110a-f05a-4511-b49a-06a2e85c44aa name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.056886715Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cb9edf2251410c340a6125ff331924ab57a72ad3166bd9e1305b6d3030284990" id=2e93926e-d482-488a-bdfd-16748f68e1ac name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.068490275Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2b208acf1f002c143fdb248dbee2517f7b6d168def4cfef1f3188939123d0dc3,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cb9edf2251410c340a6125ff331924ab57a72ad3166bd9e1305b6d3030284990],Size_:536350945,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=2e93926e-d482-488a-bdfd-16748f68e1ac name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.071863180Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cb9edf2251410c340a6125ff331924ab57a72ad3166bd9e1305b6d3030284990" id=19d8149c-c384-4645-a6c6-6443fc8eb7a4 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.072043088Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2b208acf1f002c143fdb248dbee2517f7b6d168def4cfef1f3188939123d0dc3,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cb9edf2251410c340a6125ff331924ab57a72ad3166bd9e1305b6d3030284990],Size_:536350945,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=19d8149c-c384-4645-a6c6-6443fc8eb7a4 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.072474833Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7" id=84566a46-b79b-42aa-abce-8fc9044a6cd3 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.072717205Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:6e78a44a973f2298da6f51c1418094403ee8c6c2b57bde9d3285b9e2ac2ba129,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7],Size_:397889589,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=84566a46-b79b-42aa-abce-8fc9044a6cd3 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.083527640Z" level=info msg="Creating container: openshift-machine-config-operator/machine-config-daemon-hf8f5/machine-config-daemon" id=cd10ae8b-c262-4ae4-8806-1a05cdf01028 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.083807897Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.087333938Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7" id=de99b27c-adbb-4696-acf8-c5c96ff6f213 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.087564531Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:6e78a44a973f2298da6f51c1418094403ee8c6c2b57bde9d3285b9e2ac2ba129,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7],Size_:397889589,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=de99b27c-adbb-4696-acf8-c5c96ff6f213 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.091774228Z" level=info msg="Creating container: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-driver" id=9ad7801a-70aa-4c45-aa6c-8e7c254e1761 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.091894563Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.100308352Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=e0cf5218-e82a-48b9-b699-96f3ccb269d4 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 09:16:24.176317 2171 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f88cbe5_1221_4e6f_a6c9_f22da152b43f.slice/crio-9454780ea94304660f15d0806d1edaf679bb7ba5c8879c21dffe722e3df07991.scope WatchSource:0}: Error finding container 9454780ea94304660f15d0806d1edaf679bb7ba5c8879c21dffe722e3df07991: Status 404 returned error can't find the container with id 9454780ea94304660f15d0806d1edaf679bb7ba5c8879c21dffe722e3df07991 Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.194250076Z" level=info msg="Ran pod sandbox 9454780ea94304660f15d0806d1edaf679bb7ba5c8879c21dffe722e3df07991 with infra container: openshift-dns/node-resolver-nxzr8/POD" id=e0cf5218-e82a-48b9-b699-96f3ccb269d4 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.197264495Z" level=info msg="Created container 8ab726c4f14feb975596f28a3d0f3cddfa2d5bf47960c98116692e3cbad61a17: openshift-monitoring/node-exporter-2r6nf/init-textfile" id=fa4c6480-e488-4423-bc3c-8ce3d0af81f5 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.197741021Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f93a29f8d7455a5067ecde85394835bda877174193ab693b2b010d96c8cd9183" id=13e4c0ff-9804-4596-b94e-205813e98329 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.198045848Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:9b1e063a0faa2683ff00ee1f8b324af985c1c142530a63c7f8729ec8dab01b38,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f93a29f8d7455a5067ecde85394835bda877174193ab693b2b010d96c8cd9183],Size_:480731528,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=13e4c0ff-9804-4596-b94e-205813e98329 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.200547972Z" level=info msg="Starting container: 8ab726c4f14feb975596f28a3d0f3cddfa2d5bf47960c98116692e3cbad61a17" id=dd36ff5f-395f-4a58-bab4-7b508edba3ef name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.201356833Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f93a29f8d7455a5067ecde85394835bda877174193ab693b2b010d96c8cd9183" id=78b3de8f-3cc4-40c6-a343-a37037768ccb name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.207274717Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:9b1e063a0faa2683ff00ee1f8b324af985c1c142530a63c7f8729ec8dab01b38,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f93a29f8d7455a5067ecde85394835bda877174193ab693b2b010d96c8cd9183],Size_:480731528,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=78b3de8f-3cc4-40c6-a343-a37037768ccb name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.213059363Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=44eeae78-ad0e-4247-9002-93fd13f4f2d3 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.217528253Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=8fac1130-098d-4317-b72e-20f50c13091c name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.222413305Z" level=info msg="Creating container: openshift-dns/node-resolver-nxzr8/dns-node-resolver" id=bfef57d9-5ab6-4642-a219-e5751027b382 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.222611149Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 09:16:24.252489 2171 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91265f01_9dcb_4046_961e_ad4c544224d7.slice/crio-8cf844a5c2263b21afe261a444f7b662d35e088b48ebd92b2f02703a28a05ec2.scope WatchSource:0}: Error finding container 8cf844a5c2263b21afe261a444f7b662d35e088b48ebd92b2f02703a28a05ec2: Status 404 returned error can't find the container with id 8cf844a5c2263b21afe261a444f7b662d35e088b48ebd92b2f02703a28a05ec2 Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.257693216Z" level=info msg="Ran pod sandbox 8cf844a5c2263b21afe261a444f7b662d35e088b48ebd92b2f02703a28a05ec2 with infra container: openshift-cluster-node-tuning-operator/tuned-smwj7/POD" id=8fac1130-098d-4317-b72e-20f50c13091c name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.262861871Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ca6ba83315a4882f5c6c4e01c79238e1b794b0e8bfdc2f004990f516e8e0b056" id=84403675-0a2a-4bda-94ff-865fa83736a5 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.273609024Z" level=info msg="Started container" PID=2329 containerID=8ab726c4f14feb975596f28a3d0f3cddfa2d5bf47960c98116692e3cbad61a17 description=openshift-monitoring/node-exporter-2r6nf/init-textfile id=dd36ff5f-395f-4a58-bab4-7b508edba3ef name=/runtime.v1.RuntimeService/StartContainer sandboxID=a87b4ab12cf54ed2e65342c3d9c4ea51c24387aaa25fa0e287ee0befc098a632 Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.281710693Z" level=info msg="Created container 13ea8c9b5aa383f0369dbc0f8fb24015e3ec8833108b7bfb08614c7e60ccc20b: openshift-sdn/sdn-fbccx/sdn" id=77807cab-bd17-4e36-b2b9-daf0e3f17281 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.292511054Z" level=info msg="Starting container: 13ea8c9b5aa383f0369dbc0f8fb24015e3ec8833108b7bfb08614c7e60ccc20b" id=7129667e-29a3-447d-8c06-ab1d159d985a name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.295314807Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b2f8382287975a85853941eec2c606723c2ed1c24f59fa46ebdb470b91a481f5,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ca6ba83315a4882f5c6c4e01c79238e1b794b0e8bfdc2f004990f516e8e0b056],Size_:505526574,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=84403675-0a2a-4bda-94ff-865fa83736a5 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 09:16:24.302364 2171 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0e47249_a5ff_460e_958d_53577387c154.slice/crio-01ea8cd190538ff3846bfa2fc9b101a796149e8b3c9ebd27850d03c9ae1f894d.scope WatchSource:0}: Error finding container 01ea8cd190538ff3846bfa2fc9b101a796149e8b3c9ebd27850d03c9ae1f894d: Status 404 returned error can't find the container with id 01ea8cd190538ff3846bfa2fc9b101a796149e8b3c9ebd27850d03c9ae1f894d Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.340496859Z" level=info msg="Ran pod sandbox 01ea8cd190538ff3846bfa2fc9b101a796149e8b3c9ebd27850d03c9ae1f894d with infra container: openshift-multus/multus-additional-cni-plugins-5fgxc/POD" id=44eeae78-ad0e-4247-9002-93fd13f4f2d3 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.351194941Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ca6ba83315a4882f5c6c4e01c79238e1b794b0e8bfdc2f004990f516e8e0b056" id=88b4db35-1ed7-4e04-a8c5-4bf723071f0b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.351538774Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b2f8382287975a85853941eec2c606723c2ed1c24f59fa46ebdb470b91a481f5,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ca6ba83315a4882f5c6c4e01c79238e1b794b0e8bfdc2f004990f516e8e0b056],Size_:505526574,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=88b4db35-1ed7-4e04-a8c5-4bf723071f0b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.356191833Z" level=info msg="Creating container: openshift-cluster-node-tuning-operator/tuned-smwj7/tuned" id=2c613cba-96b6-4c1b-a197-2b92adb0f1e4 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.356660139Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.357123494Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:3fc008580e4e404960b2de4020a2939ab6421bdf8f51c0b60af4ad903272c741" id=4fede8c0-62ec-4a86-9a9e-8dce4f7ebe0c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.387561718Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:6bb2a44a1f0506c18ef471cac7f7e8991d1e6b9d649ccbce50f751dd2863d06d,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:3fc008580e4e404960b2de4020a2939ab6421bdf8f51c0b60af4ad903272c741],Size_:382411122,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=4fede8c0-62ec-4a86-9a9e-8dce4f7ebe0c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.429880459Z" level=info msg="Created container f1297fbd8a37d22dca4961cbf15002572c813a558196a21c193e0a7c70ecd322: openshift-machine-config-operator/machine-config-daemon-hf8f5/machine-config-daemon" id=cd10ae8b-c262-4ae4-8806-1a05cdf01028 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.430158788Z" level=info msg="Started container" PID=2353 containerID=13ea8c9b5aa383f0369dbc0f8fb24015e3ec8833108b7bfb08614c7e60ccc20b description=openshift-sdn/sdn-fbccx/sdn id=7129667e-29a3-447d-8c06-ab1d159d985a name=/runtime.v1.RuntimeService/StartContainer sandboxID=66a629012ef7e22f6af44b09d2cda79cab33d282fd659514ee53da6e316c1302 Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.458678248Z" level=info msg="Starting container: f1297fbd8a37d22dca4961cbf15002572c813a558196a21c193e0a7c70ecd322" id=6e0fd5ee-9813-4079-9ac5-05d97c07ba1a name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.463945292Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.465912313Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:3fc008580e4e404960b2de4020a2939ab6421bdf8f51c0b60af4ad903272c741" id=35afb707-bd97-4518-a5c2-5286077d105d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.466131511Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:6bb2a44a1f0506c18ef471cac7f7e8991d1e6b9d649ccbce50f751dd2863d06d,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:3fc008580e4e404960b2de4020a2939ab6421bdf8f51c0b60af4ad903272c741],Size_:382411122,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=35afb707-bd97-4518-a5c2-5286077d105d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.486991624Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/egress-router-binary-copy" id=cac0bf3e-5271-4ad7-915f-fba47895cb3b name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.487230580Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.546615348Z" level=info msg="Started container" PID=2388 containerID=f1297fbd8a37d22dca4961cbf15002572c813a558196a21c193e0a7c70ecd322 description=openshift-machine-config-operator/machine-config-daemon-hf8f5/machine-config-daemon id=6e0fd5ee-9813-4079-9ac5-05d97c07ba1a name=/runtime.v1.RuntimeService/StartContainer sandboxID=c41ae071597e07deb9c5ea067a36e06b0f2b19fc9c9d6633270660715cffd374 Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.552951902Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=64f9368d-d2f2-48a5-ae08-045d9c8468e1 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.588961228Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=64f9368d-d2f2-48a5-ae08-045d9c8468e1 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.601243133Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.601393508Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.601488013Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.619733074Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=0c3393ea-35cd-48fe-bd97-668f44176eba name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.620170528Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=0c3393ea-35cd-48fe-bd97-668f44176eba name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.647945490Z" level=info msg="Created container 66b006e10ff01dda7aaf666ba175d1fe5ae43bbe9f6ebf670dbb4064be349276: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-driver" id=9ad7801a-70aa-4c45-aa6c-8e7c254e1761 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.690418876Z" level=info msg="Starting container: 66b006e10ff01dda7aaf666ba175d1fe5ae43bbe9f6ebf670dbb4064be349276" id=b9c3b31e-80d2-4061-9641-5c1028a05e16 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.699799384Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=2ef19a62-c4f3-415e-9e76-9b0b3b242c4a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.701777068Z" level=info msg="Creating container: openshift-sdn/sdn-fbccx/kube-rbac-proxy" id=9813aa18-9905-4d12-8df0-ab528936a180 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.701922308Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.719409908Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=2ef19a62-c4f3-415e-9e76-9b0b3b242c4a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.742680690Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=ac6202d2-2298-4f2a-9cab-72d2623bbf0c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.758871202Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.758924804Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.821725871Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=ac6202d2-2298-4f2a-9cab-72d2623bbf0c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.846155061Z" level=info msg="Creating container: openshift-machine-config-operator/machine-config-daemon-hf8f5/oauth-proxy" id=5ec3d61f-8d07-4dfe-b843-42d80646379f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.846447851Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:24.850297163Z" level=info msg="Started container" PID=2523 containerID=66b006e10ff01dda7aaf666ba175d1fe5ae43bbe9f6ebf670dbb4064be349276 description=openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-driver id=b9c3b31e-80d2-4061-9641-5c1028a05e16 name=/runtime.v1.RuntimeService/StartContainer sandboxID=821f01442b5e38ec5d5958577a6f01e03591ea0d8fc60b05f51cb69558bd0cb1 Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:24.959626 2171 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=02e0075b-c660-4dc5-8cce-00fc4ef2980f path="/var/lib/kubelet/pods/02e0075b-c660-4dc5-8cce-00fc4ef2980f/volumes" Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:24.964962 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-xw6d2" event=&{ID:659b8bb4-81b5-49bf-8127-1d326f30ba2f Type:ContainerStarted Data:080616c6ee5360407adb7a3e03767428f4727a0ee1746976a16081b2970f8d20} Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:24.993094 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerStarted Data:01ea8cd190538ff3846bfa2fc9b101a796149e8b3c9ebd27850d03c9ae1f894d} Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:24.996249 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" event=&{ID:997b7ab4-939e-465c-9c7d-4a2ebe3a797a Type:ContainerStarted Data:f1297fbd8a37d22dca4961cbf15002572c813a558196a21c193e0a7c70ecd322} Jan 05 09:16:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:24.996282 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" event=&{ID:997b7ab4-939e-465c-9c7d-4a2ebe3a797a Type:ContainerStarted Data:c41ae071597e07deb9c5ea067a36e06b0f2b19fc9c9d6633270660715cffd374} Jan 05 09:16:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:25.003016 2171 generic.go:296] "Generic (PLEG): container finished" podID=aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f containerID="8ab726c4f14feb975596f28a3d0f3cddfa2d5bf47960c98116692e3cbad61a17" exitCode=0 Jan 05 09:16:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:25.003070 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-2r6nf" event=&{ID:aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f Type:ContainerDied Data:8ab726c4f14feb975596f28a3d0f3cddfa2d5bf47960c98116692e3cbad61a17} Jan 05 09:16:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:25.004159911Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180" id=8527e2e2-28fa-4d23-b5cb-f87b25f941fa name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:25.006048778Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b4cde205857c1e0a984b38c61275a8370136432c7c84dbfbd7f4835f26bbb935,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180],Size_:297268865,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=8527e2e2-28fa-4d23-b5cb-f87b25f941fa name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:25.008543305Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180" id=e5bf7950-9b4e-4cbe-939b-1963361a9bb6 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:25.008929676Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b4cde205857c1e0a984b38c61275a8370136432c7c84dbfbd7f4835f26bbb935,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180],Size_:297268865,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=e5bf7950-9b4e-4cbe-939b-1963361a9bb6 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:25.012989412Z" level=info msg="Creating container: openshift-monitoring/node-exporter-2r6nf/node-exporter" id=bff19abf-c4a7-41fe-a6ca-7826d1de1d59 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:25.015064981Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:25.044273 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-nxzr8" event=&{ID:8f88cbe5-1221-4e6f-a6c9-f22da152b43f Type:ContainerStarted Data:9454780ea94304660f15d0806d1edaf679bb7ba5c8879c21dffe722e3df07991} Jan 05 09:16:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:25.049036 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-smwj7" event=&{ID:91265f01-9dcb-4046-961e-ad4c544224d7 Type:ContainerStarted Data:8cf844a5c2263b21afe261a444f7b662d35e088b48ebd92b2f02703a28a05ec2} Jan 05 09:16:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:25.055004 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" event=&{ID:1bb84a70-232c-4467-ac0b-647030155908 Type:ContainerStarted Data:821f01442b5e38ec5d5958577a6f01e03591ea0d8fc60b05f51cb69558bd0cb1} Jan 05 09:16:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:25.075509 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-sdn/sdn-fbccx" event=&{ID:b5d621a5-07d9-459b-8745-a991cf4a179e Type:ContainerStarted Data:13ea8c9b5aa383f0369dbc0f8fb24015e3ec8833108b7bfb08614c7e60ccc20b} Jan 05 09:16:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:25.075542 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-sdn/sdn-fbccx" event=&{ID:b5d621a5-07d9-459b-8745-a991cf4a179e Type:ContainerStarted Data:66a629012ef7e22f6af44b09d2cda79cab33d282fd659514ee53da6e316c1302} Jan 05 09:16:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:25.168236413Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:72c764b55efc147864960930ae05b7928c5f976eb76820b38997f0df7afab820" id=0938cf1c-c1e1-479a-b2e8-1dfae53a5c17 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:25.212147154Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:0bb939dc4e915f972f52c355cea411db098a9c605b2ca4bc1048447ae297391e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:72c764b55efc147864960930ae05b7928c5f976eb76820b38997f0df7afab820],Size_:293414029,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=0938cf1c-c1e1-479a-b2e8-1dfae53a5c17 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:25.254889667Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:72c764b55efc147864960930ae05b7928c5f976eb76820b38997f0df7afab820" id=805c37dc-83f1-41a9-b4b3-78af47c4419f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:25.267742452Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:0bb939dc4e915f972f52c355cea411db098a9c605b2ca4bc1048447ae297391e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:72c764b55efc147864960930ae05b7928c5f976eb76820b38997f0df7afab820],Size_:293414029,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=805c37dc-83f1-41a9-b4b3-78af47c4419f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:25.287925287Z" level=info msg="Creating container: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-node-driver-registrar" id=505e97f0-2796-4bd1-b29a-789164d2eb9f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:25.288044559Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 09:16:25.732904 2171 watcher.go:93] Error while processing event ("/sys/fs/cgroup/pids/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode13f5318_440f_4b4e_863d_2489155b9feb.slice/crio-conmon-e6c544472a3fe1c13d8c25f801c04018deedf4070a00354504f4f053617bcbec.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/pids/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode13f5318_440f_4b4e_863d_2489155b9feb.slice/crio-conmon-e6c544472a3fe1c13d8c25f801c04018deedf4070a00354504f4f053617bcbec.scope: no such file or directory Jan 05 09:16:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 09:16:25.733201 2171 container.go:488] Failed to get RecentStats("/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode13f5318_440f_4b4e_863d_2489155b9feb.slice/crio-conmon-e6c544472a3fe1c13d8c25f801c04018deedf4070a00354504f4f053617bcbec.scope") while determining the next housekeeping: unable to find data in memory cache Jan 05 09:16:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:25.810599482Z" level=info msg="Created container 0613c3c864f6a738047b48ad32d75cbac23078d274bd5c606acafbde88c5f043: openshift-cluster-node-tuning-operator/tuned-smwj7/tuned" id=2c613cba-96b6-4c1b-a197-2b92adb0f1e4 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:25.816410435Z" level=info msg="Starting container: 0613c3c864f6a738047b48ad32d75cbac23078d274bd5c606acafbde88c5f043" id=6bc0f1bd-18b3-4dce-a591-442c81d90144 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:26.116929 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" event=&{ID:1bb84a70-232c-4467-ac0b-647030155908 Type:ContainerStarted Data:66b006e10ff01dda7aaf666ba175d1fe5ae43bbe9f6ebf670dbb4064be349276} Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.116003791Z" level=info msg="Started container" PID=2570 containerID=0613c3c864f6a738047b48ad32d75cbac23078d274bd5c606acafbde88c5f043 description=openshift-cluster-node-tuning-operator/tuned-smwj7/tuned id=6bc0f1bd-18b3-4dce-a591-442c81d90144 name=/runtime.v1.RuntimeService/StartContainer sandboxID=8cf844a5c2263b21afe261a444f7b662d35e088b48ebd92b2f02703a28a05ec2 Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.305160025Z" level=info msg="Created container 54eac0f3e9474ac7129b37dd301317b64b5cf72dffa4afd2f5e25bb030a8704e: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-node-driver-registrar" id=505e97f0-2796-4bd1-b29a-789164d2eb9f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.307698998Z" level=info msg="Starting container: 54eac0f3e9474ac7129b37dd301317b64b5cf72dffa4afd2f5e25bb030a8704e" id=ee6b1022-cf98-4923-9674-930363dadbea name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.313723810Z" level=info msg="Created container b84dab7a298b3be0b8ad3e96ce486984126dd9280a93922e23840d85810b4733: openshift-sdn/sdn-fbccx/kube-rbac-proxy" id=9813aa18-9905-4d12-8df0-ab528936a180 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.317423551Z" level=info msg="Starting container: b84dab7a298b3be0b8ad3e96ce486984126dd9280a93922e23840d85810b4733" id=8442924f-fcf1-471f-bc1f-ebb24d94c0db name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.328524954Z" level=info msg="Created container ba11f646f8ef5ee22d313d2097fb88998b3c4d4bf2a1eac1162cf99c7a3730be: openshift-monitoring/node-exporter-2r6nf/node-exporter" id=bff19abf-c4a7-41fe-a6ca-7826d1de1d59 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.330585905Z" level=info msg="Starting container: ba11f646f8ef5ee22d313d2097fb88998b3c4d4bf2a1eac1162cf99c7a3730be" id=fad52617-15d3-4d88-90d7-1f624c54569c name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.358577197Z" level=info msg="Created container 9c6af66a2c255e371dd043bab4278fde4055eb8fbdef8d4d771a8ba43dfecec6: openshift-machine-config-operator/machine-config-daemon-hf8f5/oauth-proxy" id=5ec3d61f-8d07-4dfe-b843-42d80646379f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.359210457Z" level=info msg="Starting container: 9c6af66a2c255e371dd043bab4278fde4055eb8fbdef8d4d771a8ba43dfecec6" id=134f32f9-6366-4381-8db2-eb9e7cfea6b5 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.404218752Z" level=info msg="Created container d9239b02fd1e57f00a1306dc785edd6cd0f57d1ea5e97daf35cd915076c0d8b5: openshift-image-registry/node-ca-xw6d2/node-ca" id=616715eb-122a-4891-90c8-a9935a97c31e name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.404838409Z" level=info msg="Starting container: d9239b02fd1e57f00a1306dc785edd6cd0f57d1ea5e97daf35cd915076c0d8b5" id=4c182921-a5dd-4054-9065-e289b2543d7b name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.406941959Z" level=info msg="Started container" PID=2853 containerID=b84dab7a298b3be0b8ad3e96ce486984126dd9280a93922e23840d85810b4733 description=openshift-sdn/sdn-fbccx/kube-rbac-proxy id=8442924f-fcf1-471f-bc1f-ebb24d94c0db name=/runtime.v1.RuntimeService/StartContainer sandboxID=66a629012ef7e22f6af44b09d2cda79cab33d282fd659514ee53da6e316c1302 Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.429840718Z" level=info msg="Started container" PID=2828 containerID=ba11f646f8ef5ee22d313d2097fb88998b3c4d4bf2a1eac1162cf99c7a3730be description=openshift-monitoring/node-exporter-2r6nf/node-exporter id=fad52617-15d3-4d88-90d7-1f624c54569c name=/runtime.v1.RuntimeService/StartContainer sandboxID=a87b4ab12cf54ed2e65342c3d9c4ea51c24387aaa25fa0e287ee0befc098a632 Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.463811950Z" level=info msg="Started container" PID=2851 containerID=54eac0f3e9474ac7129b37dd301317b64b5cf72dffa4afd2f5e25bb030a8704e description=openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-node-driver-registrar id=ee6b1022-cf98-4923-9674-930363dadbea name=/runtime.v1.RuntimeService/StartContainer sandboxID=821f01442b5e38ec5d5958577a6f01e03591ea0d8fc60b05f51cb69558bd0cb1 Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.466770114Z" level=info msg="Created container e6c544472a3fe1c13d8c25f801c04018deedf4070a00354504f4f053617bcbec: openshift-multus/multus-twl8f/kube-multus" id=69821c55-1902-4c8a-9586-ea3ed4bb6fb8 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.470209880Z" level=info msg="Starting container: e6c544472a3fe1c13d8c25f801c04018deedf4070a00354504f4f053617bcbec" id=21a83530-dc6e-41a3-a7fd-23ad2b34891a name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.504640150Z" level=info msg="Started container" PID=2831 containerID=9c6af66a2c255e371dd043bab4278fde4055eb8fbdef8d4d771a8ba43dfecec6 description=openshift-machine-config-operator/machine-config-daemon-hf8f5/oauth-proxy id=134f32f9-6366-4381-8db2-eb9e7cfea6b5 name=/runtime.v1.RuntimeService/StartContainer sandboxID=c41ae071597e07deb9c5ea067a36e06b0f2b19fc9c9d6633270660715cffd374 Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.531255680Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=f7fa0171-b911-463b-bbf5-1ca7720f874a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.531637896Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=f7fa0171-b911-463b-bbf5-1ca7720f874a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.550586559Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=7ceeb370-78a6-41c7-8956-7420d786f677 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.554994577Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=7ceeb370-78a6-41c7-8956-7420d786f677 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.561058041Z" level=info msg="Creating container: openshift-monitoring/node-exporter-2r6nf/kube-rbac-proxy" id=16b613fd-2752-4273-b89a-4220c415ea45 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.561922571Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.574525656Z" level=info msg="Created container e050307cf2a700b4a984c2345a643e84a5f9d0e6a3d741aec158e92ce185de1b: openshift-multus/multus-additional-cni-plugins-5fgxc/egress-router-binary-copy" id=cac0bf3e-5271-4ad7-915f-fba47895cb3b name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.575303578Z" level=info msg="Starting container: e050307cf2a700b4a984c2345a643e84a5f9d0e6a3d741aec158e92ce185de1b" id=3675821d-8c09-446b-8df8-bfa9ae27a545 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.583793538Z" level=info msg="Started container" PID=2830 containerID=d9239b02fd1e57f00a1306dc785edd6cd0f57d1ea5e97daf35cd915076c0d8b5 description=openshift-image-registry/node-ca-xw6d2/node-ca id=4c182921-a5dd-4054-9065-e289b2543d7b name=/runtime.v1.RuntimeService/StartContainer sandboxID=080616c6ee5360407adb7a3e03767428f4727a0ee1746976a16081b2970f8d20 Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.663761299Z" level=info msg="Started container" PID=2915 containerID=e6c544472a3fe1c13d8c25f801c04018deedf4070a00354504f4f053617bcbec description=openshift-multus/multus-twl8f/kube-multus id=21a83530-dc6e-41a3-a7fd-23ad2b34891a name=/runtime.v1.RuntimeService/StartContainer sandboxID=d04b2e4ec3af58e65ad07d97f2565affd7e50b26af21f25245500b2278268045 Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.664454444Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ea25c2f3749dee6981359b4655a9acfcd4d95547066ccdf328354b548da74932" id=28ef9a9c-663b-4436-9b34-d0655933fc2e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.664765899Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:58a5591f9334b70902e27f9a172f3e141b8f07a879507b850a76c2499014eec9,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ea25c2f3749dee6981359b4655a9acfcd4d95547066ccdf328354b548da74932],Size_:291396124,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=28ef9a9c-663b-4436-9b34-d0655933fc2e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.711171848Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ea25c2f3749dee6981359b4655a9acfcd4d95547066ccdf328354b548da74932" id=ff013331-4229-4437-8f71-fdf281213536 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.711560571Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:58a5591f9334b70902e27f9a172f3e141b8f07a879507b850a76c2499014eec9,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ea25c2f3749dee6981359b4655a9acfcd4d95547066ccdf328354b548da74932],Size_:291396124,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=ff013331-4229-4437-8f71-fdf281213536 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.718076877Z" level=info msg="Creating container: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-liveness-probe" id=8e14c8e1-795d-43b2-a7f2-426095cd9b31 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.718298362Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.768550585Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_29f4e067-f126-4410-a8a7-af4f8cb93f1a\"" Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.840941981Z" level=info msg="Started container" PID=2909 containerID=e050307cf2a700b4a984c2345a643e84a5f9d0e6a3d741aec158e92ce185de1b description=openshift-multus/multus-additional-cni-plugins-5fgxc/egress-router-binary-copy id=3675821d-8c09-446b-8df8-bfa9ae27a545 name=/runtime.v1.RuntimeService/StartContainer sandboxID=01ea8cd190538ff3846bfa2fc9b101a796149e8b3c9ebd27850d03c9ae1f894d Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.869091250Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.869256385Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.884648516Z" level=info msg="Created container 33ba7961e207da54e57a0776bd2ca3df719a3f5577d51f8316e4e896a106bc0e: openshift-dns/node-resolver-nxzr8/dns-node-resolver" id=bfef57d9-5ab6-4642-a219-e5751027b382 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.885892790Z" level=info msg="Starting container: 33ba7961e207da54e57a0776bd2ca3df719a3f5577d51f8316e4e896a106bc0e" id=bd7935ae-8b0f-426a-9a50-701d9dd0cf07 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.950862862Z" level=info msg="Started container" PID=2990 containerID=33ba7961e207da54e57a0776bd2ca3df719a3f5577d51f8316e4e896a106bc0e description=openshift-dns/node-resolver-nxzr8/dns-node-resolver id=bd7935ae-8b0f-426a-9a50-701d9dd0cf07 name=/runtime.v1.RuntimeService/StartContainer sandboxID=9454780ea94304660f15d0806d1edaf679bb7ba5c8879c21dffe722e3df07991 Jan 05 09:16:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:26.974876632Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_2d0eba4c-98e0-44e7-88b8-4b046675590e\"" Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:27.008872343Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:27.009040317Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:27.054797233Z" level=info msg="Created container 5bf3f7a5853fafbb40a74f57cc802e4b123a92fe403467c75414699f5fb683c7: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-liveness-probe" id=8e14c8e1-795d-43b2-a7f2-426095cd9b31 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:27.055358390Z" level=info msg="Starting container: 5bf3f7a5853fafbb40a74f57cc802e4b123a92fe403467c75414699f5fb683c7" id=61ec9861-5328-4dfc-b7e9-946dc9f0c5f7 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:27.073485113Z" level=info msg="Started container" PID=3166 containerID=5bf3f7a5853fafbb40a74f57cc802e4b123a92fe403467c75414699f5fb683c7 description=openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-liveness-probe id=61ec9861-5328-4dfc-b7e9-946dc9f0c5f7 name=/runtime.v1.RuntimeService/StartContainer sandboxID=821f01442b5e38ec5d5958577a6f01e03591ea0d8fc60b05f51cb69558bd0cb1 Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:27.120172 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-nxzr8" event=&{ID:8f88cbe5-1221-4e6f-a6c9-f22da152b43f Type:ContainerStarted Data:33ba7961e207da54e57a0776bd2ca3df719a3f5577d51f8316e4e896a106bc0e} Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:27.126783 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-smwj7" event=&{ID:91265f01-9dcb-4046-961e-ad4c544224d7 Type:ContainerStarted Data:0613c3c864f6a738047b48ad32d75cbac23078d274bd5c606acafbde88c5f043} Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:27.130881 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" event=&{ID:1bb84a70-232c-4467-ac0b-647030155908 Type:ContainerStarted Data:5bf3f7a5853fafbb40a74f57cc802e4b123a92fe403467c75414699f5fb683c7} Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:27.130923 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" event=&{ID:1bb84a70-232c-4467-ac0b-647030155908 Type:ContainerStarted Data:54eac0f3e9474ac7129b37dd301317b64b5cf72dffa4afd2f5e25bb030a8704e} Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:27.136552 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-2r6nf" event=&{ID:aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f Type:ContainerStarted Data:ba11f646f8ef5ee22d313d2097fb88998b3c4d4bf2a1eac1162cf99c7a3730be} Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:27.140439 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-twl8f" event=&{ID:e13f5318-440f-4b4e-863d-2489155b9feb Type:ContainerStarted Data:e6c544472a3fe1c13d8c25f801c04018deedf4070a00354504f4f053617bcbec} Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:27.144518 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-xw6d2" event=&{ID:659b8bb4-81b5-49bf-8127-1d326f30ba2f Type:ContainerStarted Data:d9239b02fd1e57f00a1306dc785edd6cd0f57d1ea5e97daf35cd915076c0d8b5} Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:27.149210 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-sdn/sdn-fbccx" event=&{ID:b5d621a5-07d9-459b-8745-a991cf4a179e Type:ContainerStarted Data:b84dab7a298b3be0b8ad3e96ce486984126dd9280a93922e23840d85810b4733} Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:27.160854 2171 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-sdn/sdn-fbccx" Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:27.160932 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerStarted Data:e050307cf2a700b4a984c2345a643e84a5f9d0e6a3d741aec158e92ce185de1b} Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:27.164875 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" event=&{ID:997b7ab4-939e-465c-9c7d-4a2ebe3a797a Type:ContainerStarted Data:9c6af66a2c255e371dd043bab4278fde4055eb8fbdef8d4d771a8ba43dfecec6} Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:27.393637288Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/egress-router\"" Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:27.418111600Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:27.418162761Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:27.439469079Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/multus\"" Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:27.475380844Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:27.475438938Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:27.475458615Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_2d0eba4c-98e0-44e7-88b8-4b046675590e\"" Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:27.475482775Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_29f4e067-f126-4410-a8a7-af4f8cb93f1a\"" Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:27.553743926Z" level=info msg="Created container af19bb0973727cee70da4a13aa5e96309c2c728a8412370d3f1bd73539a45ad5: openshift-monitoring/node-exporter-2r6nf/kube-rbac-proxy" id=16b613fd-2752-4273-b89a-4220c415ea45 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:27.554653415Z" level=info msg="Starting container: af19bb0973727cee70da4a13aa5e96309c2c728a8412370d3f1bd73539a45ad5" id=a4c3499e-d0fb-4881-be96-781946f7c14d name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:27.566294831Z" level=info msg="Started container" PID=3246 containerID=af19bb0973727cee70da4a13aa5e96309c2c728a8412370d3f1bd73539a45ad5 description=openshift-monitoring/node-exporter-2r6nf/kube-rbac-proxy id=a4c3499e-d0fb-4881-be96-781946f7c14d name=/runtime.v1.RuntimeService/StartContainer sandboxID=a87b4ab12cf54ed2e65342c3d9c4ea51c24387aaa25fa0e287ee0befc098a632 Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:27.770257335Z" level=info msg="CNI monitoring event REMOVE \"/etc/kubernetes/cni/net.d/00-multus.conf\"" Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:27.786201831Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:27.786371289Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:27.786480207Z" level=info msg="CNI monitoring event CREATE \"/etc/kubernetes/cni/net.d/00-multus.conf\"" Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:27.794737530Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:27.794779112Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:27.794796889Z" level=info msg="CNI monitoring event WRITE \"/etc/kubernetes/cni/net.d/00-multus.conf\"" Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:27.802059782Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:27.802078658Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:27.802092607Z" level=info msg="CNI monitoring event CHMOD \"/etc/kubernetes/cni/net.d/00-multus.conf\"" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:28.168473 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-2r6nf" event=&{ID:aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f Type:ContainerStarted Data:af19bb0973727cee70da4a13aa5e96309c2c728a8412370d3f1bd73539a45ad5} Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:28.172062 2171 generic.go:296] "Generic (PLEG): container finished" podID=c0e47249-a5ff-460e-958d-53577387c154 containerID="e050307cf2a700b4a984c2345a643e84a5f9d0e6a3d741aec158e92ce185de1b" exitCode=0 Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.174040856Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d5fc161170fcf68900b2694039803a2e7f1b31eb913566139a41f91851f166d7" id=a49f1b6c-35c7-48ab-804f-3860a844e7ad name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.174305338Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:852cc803756e640141d6c3eb3e6607f2b0f7076ea77f9190b1f490288f40b3d8,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d5fc161170fcf68900b2694039803a2e7f1b31eb913566139a41f91851f166d7],Size_:574516705,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=a49f1b6c-35c7-48ab-804f-3860a844e7ad name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:28.172840 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerDied Data:e050307cf2a700b4a984c2345a643e84a5f9d0e6a3d741aec158e92ce185de1b} Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.187677806Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d5fc161170fcf68900b2694039803a2e7f1b31eb913566139a41f91851f166d7" id=0da53c25-1392-4fd0-8e89-aca6b2f6c46b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.187903606Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:852cc803756e640141d6c3eb3e6607f2b0f7076ea77f9190b1f490288f40b3d8,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d5fc161170fcf68900b2694039803a2e7f1b31eb913566139a41f91851f166d7],Size_:574516705,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=0da53c25-1392-4fd0-8e89-aca6b2f6c46b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.191363831Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/cni-plugins" id=979086fe-f633-4780-b6a4-5974ea061f90 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.191505616Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: 2023-01-05T09:16:28Z [verbose] Del: openshift-dns:dns-default-2pfzf:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.355049125Z" level=info msg="Successfully cleaned up network for pod 81b981c00eed3ce8e7b961151a721e4be9ef2eef83f4433fad233b19e63776bf" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.355154092Z" level=info msg="cleanup sandbox network" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.355239358Z" level=info msg="Successfully cleaned up network for pod f160f14931d6ef807fb1308f5ac27f062e39c5cad361cedf59e27a46a8b3bd00" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.355298237Z" level=info msg="cleanup sandbox network" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.355361190Z" level=info msg="Successfully cleaned up network for pod 63ad0ecbd8ba6a412b2c14895d6a13d0ead1d3ee31d0f15597f1df586d1d96e9" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.355450072Z" level=info msg="cleanup sandbox network" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.356636064Z" level=info msg="Got pod network &{Name:dns-default-2pfzf Namespace:openshift-dns ID:63f26fe630a86618ccb6926337ae0f77635a06eec51dd1cb5b91c86eaefd519e UID:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 NetNS:/var/run/netns/e6fe9a72-a6aa-41d5-86f7-be1f22acf516 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.356808194Z" level=info msg="Adding pod openshift-dns_dns-default-2pfzf to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.646495252Z" level=info msg="Created container d8b02268262787145aa6775f3263b7cf85fa31363ace2001926a3c4cf2c4f250: openshift-multus/multus-additional-cni-plugins-5fgxc/cni-plugins" id=979086fe-f633-4780-b6a4-5974ea061f90 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.648362448Z" level=info msg="Starting container: d8b02268262787145aa6775f3263b7cf85fa31363ace2001926a3c4cf2c4f250" id=79091f92-67c7-4f74-960a-d7beb4e45c29 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.659389229Z" level=info msg="Started container" PID=3620 containerID=d8b02268262787145aa6775f3263b7cf85fa31363ace2001926a3c4cf2c4f250 description=openshift-multus/multus-additional-cni-plugins-5fgxc/cni-plugins id=79091f92-67c7-4f74-960a-d7beb4e45c29 name=/runtime.v1.RuntimeService/StartContainer sandboxID=01ea8cd190538ff3846bfa2fc9b101a796149e8b3c9ebd27850d03c9ae1f894d Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.677327275Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_f8ba4a76-7347-4d10-8e03-985118e6cdbb\"" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.694507992Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.694645170Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.859562839Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/bandwidth\"" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.871790582Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.871814004Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.871830677Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/bridge\"" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.903784799Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.903916384Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.903999155Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/dhcp\"" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.930609445Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.930914895Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.931012336Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/firewall\"" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.955414513Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.955580348Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.955664794Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/host-device\"" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.970660676Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.970791671Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.970871397Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/host-local\"" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.979480874Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.979499772Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.979514776Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/ipvlan\"" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.989931330Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.989951815Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.989979143Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/loopback\"" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.997331069Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.997349614Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:28.997365739Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/macvlan\"" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.004621372Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.004638837Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.004653302Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/portmap\"" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.012985211Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.013004544Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.013018323Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/ptp\"" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.020708363Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.020725637Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.020739344Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/sbr\"" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.027797490Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.027815382Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.027831059Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/static\"" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.035048813Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.035066285Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.035079464Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/tuning\"" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.042212146Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.042229258Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.042242281Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/vlan\"" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.049155107Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.049172469Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.049187337Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/vrf\"" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.056615047Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.056633206Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.056651809Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_f8ba4a76-7347-4d10-8e03-985118e6cdbb\"" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:29.177228 2171 generic.go:296] "Generic (PLEG): container finished" podID=c0e47249-a5ff-460e-958d-53577387c154 containerID="d8b02268262787145aa6775f3263b7cf85fa31363ace2001926a3c4cf2c4f250" exitCode=0 Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:29.177438 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerDied Data:d8b02268262787145aa6775f3263b7cf85fa31363ace2001926a3c4cf2c4f250} Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.179850689Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7ad3b9824ea195f38b6346262f0c609415c487ffaeb554725621a8327266da0f" id=21f05329-55fd-47ef-9119-448a7969b5e3 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.180271393Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:4c7cce1e79d06b3bc147fc59bd0a72991be97e767193f661db48974f3bb9ac30,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7ad3b9824ea195f38b6346262f0c609415c487ffaeb554725621a8327266da0f],Size_:336119748,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=21f05329-55fd-47ef-9119-448a7969b5e3 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.182283274Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7ad3b9824ea195f38b6346262f0c609415c487ffaeb554725621a8327266da0f" id=0aa9a626-02d6-493d-a0f5-b28c60742639 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.182616953Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:4c7cce1e79d06b3bc147fc59bd0a72991be97e767193f661db48974f3bb9ac30,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7ad3b9824ea195f38b6346262f0c609415c487ffaeb554725621a8327266da0f],Size_:336119748,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=0aa9a626-02d6-493d-a0f5-b28c60742639 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.183366598Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/bond-cni-plugin" id=928f3f62-db1e-4953-b08c-46f6514a51e9 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.183515457Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:29.247945 2171 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-sdn/sdn-fbccx" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.537127892Z" level=info msg="Created container 7578d7bf6cd2d2172abb6b188e73d8fb8ff6b5b8946d9c2da442ed1cc948ab6f: openshift-multus/multus-additional-cni-plugins-5fgxc/bond-cni-plugin" id=928f3f62-db1e-4953-b08c-46f6514a51e9 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.537559171Z" level=info msg="Starting container: 7578d7bf6cd2d2172abb6b188e73d8fb8ff6b5b8946d9c2da442ed1cc948ab6f" id=f3828358-3c35-4d65-8298-964b891686e3 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.547151483Z" level=info msg="Started container" PID=3795 containerID=7578d7bf6cd2d2172abb6b188e73d8fb8ff6b5b8946d9c2da442ed1cc948ab6f description=openshift-multus/multus-additional-cni-plugins-5fgxc/bond-cni-plugin id=f3828358-3c35-4d65-8298-964b891686e3 name=/runtime.v1.RuntimeService/StartContainer sandboxID=01ea8cd190538ff3846bfa2fc9b101a796149e8b3c9ebd27850d03c9ae1f894d Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.572160055Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_d2190c02-d583-485f-ac28-e0ce414ddab8\"" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.590949899Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.591100187Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.601739002Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/bond\"" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.612619175Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.612638938Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.612655366Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_d2190c02-d583-485f-ac28-e0ce414ddab8\"" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:29.693520 2171 plugin_watcher.go:203] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/diskplugin.csi.alibabacloud.com-reg.sock" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:29.752944 2171 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-network-diagnostics/network-check-target-xmq2g] Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: 2023-01-05T09:16:29Z [verbose] Add: openshift-network-diagnostics:network-check-target-xmq2g:f364a949-09d7-4cfb-83ff-e532b822a557:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/4e35a0fe-47dd-4c20-9ad8-c4b066c02651"}],"ips":[{"version":"4","interface":0,"address":"10.129.2.3/23","gateway":"10.129.2.1"}],"dns":{}} Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: I0105 09:16:29.737025 2296 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-network-diagnostics", Name:"network-check-target-xmq2g", UID:"f364a949-09d7-4cfb-83ff-e532b822a557", APIVersion:"v1", ResourceVersion:"79293", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.129.2.3/23] from openshift-sdn Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.754202365Z" level=info msg="Got pod network &{Name:network-check-target-xmq2g Namespace:openshift-network-diagnostics ID:00fb6557a9939b9a2bc7c25ec8cc1770b954010cd506656807f39f57849767a8 UID:f364a949-09d7-4cfb-83ff-e532b822a557 NetNS:/var/run/netns/4e35a0fe-47dd-4c20-9ad8-c4b066c02651 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.754348603Z" level=info msg="Checking pod openshift-network-diagnostics_network-check-target-xmq2g for CNI network multus-cni-network (type=multus)" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.754497752Z" level=info msg="Got pod network &{Name:network-check-target-xmq2g Namespace:openshift-network-diagnostics ID:e1cf6641b5c3c13478f683a4318f93d2bf7b7d9376939506acf8a588476f10eb UID:f364a949-09d7-4cfb-83ff-e532b822a557 NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.754629684Z" level=info msg="Deleting pod openshift-network-diagnostics_network-check-target-xmq2g from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 09:16:29.757780 2171 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf364a949_09d7_4cfb_83ff_e532b822a557.slice/crio-00fb6557a9939b9a2bc7c25ec8cc1770b954010cd506656807f39f57849767a8.scope WatchSource:0}: Error finding container 00fb6557a9939b9a2bc7c25ec8cc1770b954010cd506656807f39f57849767a8: Status 404 returned error can't find the container with id 00fb6557a9939b9a2bc7c25ec8cc1770b954010cd506656807f39f57849767a8 Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.768515230Z" level=info msg="Ran pod sandbox 00fb6557a9939b9a2bc7c25ec8cc1770b954010cd506656807f39f57849767a8 with infra container: openshift-network-diagnostics/network-check-target-xmq2g/POD" id=05d8b8df-127d-4078-a91a-6443ff5ac6fb name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.770362651Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7" id=e63e0bb4-c6d1-45e8-999a-d7a2853dd029 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.770705240Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:4cb41c73a232ec4401c5db32cdb523810dc12ac605cdc6a9ce388e90930eb6c4,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7],Size_:515955107,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=e63e0bb4-c6d1-45e8-999a-d7a2853dd029 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.772301656Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7" id=a98458b6-6e06-4b35-8f0b-5a82aa087f71 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.772620766Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:4cb41c73a232ec4401c5db32cdb523810dc12ac605cdc6a9ce388e90930eb6c4,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7],Size_:515955107,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=a98458b6-6e06-4b35-8f0b-5a82aa087f71 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.773531770Z" level=info msg="Creating container: openshift-network-diagnostics/network-check-target-xmq2g/network-check-target-container" id=aab51689-5a71-4c92-a629-26f0a6766710 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.773639312Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:29.848287 2171 reconciler.go:164] "OperationExecutor.RegisterPlugin started" plugin={SocketPath:/var/lib/kubelet/plugins_registry/diskplugin.csi.alibabacloud.com-reg.sock Timestamp:2023-01-05 09:16:29.693541288 +0000 UTC m=+7.543398476 Handler: Name:} Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:29.853224 2171 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-ingress-canary/ingress-canary-8xrbh] Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:29.856234 2171 csi_plugin.go:99] kubernetes.io/csi: Trying to validate a new CSI Driver with name: diskplugin.csi.alibabacloud.com endpoint: /var/lib/kubelet/plugins/diskplugin.csi.alibabacloud.com/csi.sock versions: 1.0.0 Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:29.856354 2171 csi_plugin.go:112] kubernetes.io/csi: Register new plugin with name: diskplugin.csi.alibabacloud.com at endpoint: /var/lib/kubelet/plugins/diskplugin.csi.alibabacloud.com/csi.sock Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: 2023-01-05T09:16:29Z [verbose] Add: openshift-ingress-canary:ingress-canary-8xrbh:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/185ab30d-a13f-46e2-a94f-58153e8799e2"}],"ips":[{"version":"4","interface":0,"address":"10.129.2.4/23","gateway":"10.129.2.1"}],"dns":{}} Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: I0105 09:16:29.837967 2305 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-ingress-canary", Name:"ingress-canary-8xrbh", UID:"e04a47b4-aa8e-4d4e-9d45-aa8037dcb748", APIVersion:"v1", ResourceVersion:"79313", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.129.2.4/23] from openshift-sdn Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.862898236Z" level=info msg="Got pod network &{Name:ingress-canary-8xrbh Namespace:openshift-ingress-canary ID:16c0a5a6c5f7e9eecc5461c963ba5b26d7e13b2c8571603f6e05ef86ce3505e3 UID:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748 NetNS:/var/run/netns/185ab30d-a13f-46e2-a94f-58153e8799e2 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.863058181Z" level=info msg="Checking pod openshift-ingress-canary_ingress-canary-8xrbh for CNI network multus-cni-network (type=multus)" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 09:16:29.867550 2171 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode04a47b4_aa8e_4d4e_9d45_aa8037dcb748.slice/crio-16c0a5a6c5f7e9eecc5461c963ba5b26d7e13b2c8571603f6e05ef86ce3505e3.scope WatchSource:0}: Error finding container 16c0a5a6c5f7e9eecc5461c963ba5b26d7e13b2c8571603f6e05ef86ce3505e3: Status 404 returned error can't find the container with id 16c0a5a6c5f7e9eecc5461c963ba5b26d7e13b2c8571603f6e05ef86ce3505e3 Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: 2023-01-05T09:16:29Z [verbose] Del: openshift-network-diagnostics:network-check-target-xmq2g:f364a949-09d7-4cfb-83ff-e532b822a557:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.877043595Z" level=info msg="Successfully cleaned up network for pod e1cf6641b5c3c13478f683a4318f93d2bf7b7d9376939506acf8a588476f10eb" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.877202504Z" level=info msg="cleanup sandbox network" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.878238449Z" level=info msg="Ran pod sandbox 16c0a5a6c5f7e9eecc5461c963ba5b26d7e13b2c8571603f6e05ef86ce3505e3 with infra container: openshift-ingress-canary/ingress-canary-8xrbh/POD" id=5b3b3170-533a-4d99-9f5e-955d1e695c6f name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.879531426Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:5974989cda2655deb3052d601e7406eaf3caee0e7999a6eff251760aac19c199" id=7c4785f4-f9b9-4bce-92b6-30e5fe612e15 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.879855703Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:c34c97b229e46faa94bab6ad2559fd683bd036df06031c9097dd27f7daaf96c7,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:5974989cda2655deb3052d601e7406eaf3caee0e7999a6eff251760aac19c199],Size_:417780553,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=7c4785f4-f9b9-4bce-92b6-30e5fe612e15 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.887572736Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:5974989cda2655deb3052d601e7406eaf3caee0e7999a6eff251760aac19c199" id=0e3574f4-5240-4c8a-9d9e-4d391a841c3d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.887835208Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:c34c97b229e46faa94bab6ad2559fd683bd036df06031c9097dd27f7daaf96c7,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:5974989cda2655deb3052d601e7406eaf3caee0e7999a6eff251760aac19c199],Size_:417780553,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=0e3574f4-5240-4c8a-9d9e-4d391a841c3d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.888700275Z" level=info msg="Creating container: openshift-ingress-canary/ingress-canary-8xrbh/serve-healthcheck-canary" id=7fd17280-770c-49da-b385-c91224c7d82d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.888821306Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.978555838Z" level=info msg="Created container ed965039c2d980207be500acfdc76412b71aa85c7c7bd8c0df6174dd65016a5d: openshift-network-diagnostics/network-check-target-xmq2g/network-check-target-container" id=aab51689-5a71-4c92-a629-26f0a6766710 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:29.979421467Z" level=info msg="Starting container: ed965039c2d980207be500acfdc76412b71aa85c7c7bd8c0df6174dd65016a5d" id=9432f034-9c5f-463c-a6d7-8a10e687259b name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.000139680Z" level=info msg="Started container" PID=3904 containerID=ed965039c2d980207be500acfdc76412b71aa85c7c7bd8c0df6174dd65016a5d description=openshift-network-diagnostics/network-check-target-xmq2g/network-check-target-container id=9432f034-9c5f-463c-a6d7-8a10e687259b name=/runtime.v1.RuntimeService/StartContainer sandboxID=00fb6557a9939b9a2bc7c25ec8cc1770b954010cd506656807f39f57849767a8 Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:30.041583 2171 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-multus/network-metrics-daemon-hq594] Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: 2023-01-05T09:16:30Z [verbose] Add: openshift-multus:network-metrics-daemon-hq594:50152296-9489-4fa8-aa42-7858debe1c08:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/74a8e5e6-0f86-443d-a641-ea34c5ca1020"}],"ips":[{"version":"4","interface":0,"address":"10.129.2.2/23","gateway":"10.129.2.1"}],"dns":{}} Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: I0105 09:16:30.023975 2286 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-multus", Name:"network-metrics-daemon-hq594", UID:"50152296-9489-4fa8-aa42-7858debe1c08", APIVersion:"v1", ResourceVersion:"79309", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.129.2.2/23] from openshift-sdn Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.045735840Z" level=info msg="Got pod network &{Name:network-metrics-daemon-hq594 Namespace:openshift-multus ID:8e2d55780ef7738a05fee323e5c71bd87390c87bd4b55a419d3d7362eb940075 UID:50152296-9489-4fa8-aa42-7858debe1c08 NetNS:/var/run/netns/74a8e5e6-0f86-443d-a641-ea34c5ca1020 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.045972449Z" level=info msg="Checking pod openshift-multus_network-metrics-daemon-hq594 for CNI network multus-cni-network (type=multus)" Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.047381728Z" level=info msg="Got pod network &{Name:network-metrics-daemon-hq594 Namespace:openshift-multus ID:5fc6d6ddb9ba864a50f6934237ca81ca5c56501c56049bb7190d5ac76187ea49 UID:50152296-9489-4fa8-aa42-7858debe1c08 NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.047631163Z" level=info msg="Deleting pod openshift-multus_network-metrics-daemon-hq594 from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 09:16:30.059916 2171 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50152296_9489_4fa8_aa42_7858debe1c08.slice/crio-8e2d55780ef7738a05fee323e5c71bd87390c87bd4b55a419d3d7362eb940075.scope WatchSource:0}: Error finding container 8e2d55780ef7738a05fee323e5c71bd87390c87bd4b55a419d3d7362eb940075: Status 404 returned error can't find the container with id 8e2d55780ef7738a05fee323e5c71bd87390c87bd4b55a419d3d7362eb940075 Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.066233812Z" level=info msg="Ran pod sandbox 8e2d55780ef7738a05fee323e5c71bd87390c87bd4b55a419d3d7362eb940075 with infra container: openshift-multus/network-metrics-daemon-hq594/POD" id=8baf2240-92cb-468d-a2fc-59548f7ca737 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.067154415Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9e33552d1aaea9cf6a301fa438cfc8305826c5ddb58ac8437d884e55e71c6e44" id=eb7d91bd-b2c1-4da2-903c-f29c265e2e99 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.067477619Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:7b87fc786118817b8913f28a61e01db37b08380f754bbaacf425508b208632a0,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9e33552d1aaea9cf6a301fa438cfc8305826c5ddb58ac8437d884e55e71c6e44],Size_:312894682,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=eb7d91bd-b2c1-4da2-903c-f29c265e2e99 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.069728243Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9e33552d1aaea9cf6a301fa438cfc8305826c5ddb58ac8437d884e55e71c6e44" id=7deae5a4-4234-4f91-8db2-5e9539f2aa41 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.069944398Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:7b87fc786118817b8913f28a61e01db37b08380f754bbaacf425508b208632a0,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9e33552d1aaea9cf6a301fa438cfc8305826c5ddb58ac8437d884e55e71c6e44],Size_:312894682,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=7deae5a4-4234-4f91-8db2-5e9539f2aa41 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.070824361Z" level=info msg="Creating container: openshift-multus/network-metrics-daemon-hq594/network-metrics-daemon" id=5ed7d158-dc25-4cd1-a727-400cc79d752d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.071018475Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.079167535Z" level=info msg="Created container 5ef6063b2be06818415c31d58e69b088971596d2dfdc9a42d50d5c9b5992d8e2: openshift-ingress-canary/ingress-canary-8xrbh/serve-healthcheck-canary" id=7fd17280-770c-49da-b385-c91224c7d82d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.079604427Z" level=info msg="Starting container: 5ef6063b2be06818415c31d58e69b088971596d2dfdc9a42d50d5c9b5992d8e2" id=9dd872ff-df4a-472c-b93c-40ca9fb74f3c name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.103269798Z" level=info msg="Started container" PID=3926 containerID=5ef6063b2be06818415c31d58e69b088971596d2dfdc9a42d50d5c9b5992d8e2 description=openshift-ingress-canary/ingress-canary-8xrbh/serve-healthcheck-canary id=9dd872ff-df4a-472c-b93c-40ca9fb74f3c name=/runtime.v1.RuntimeService/StartContainer sandboxID=16c0a5a6c5f7e9eecc5461c963ba5b26d7e13b2c8571603f6e05ef86ce3505e3 Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: 2023-01-05T09:16:30Z [verbose] Del: openshift-multus:network-metrics-daemon-hq594:50152296-9489-4fa8-aa42-7858debe1c08:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.139376171Z" level=info msg="Successfully cleaned up network for pod 5fc6d6ddb9ba864a50f6934237ca81ca5c56501c56049bb7190d5ac76187ea49" Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.139534165Z" level=info msg="cleanup sandbox network" Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.139625217Z" level=info msg="Successfully cleaned up network for pod d611273f3568d300e64d6a51cb6de91bcf31c3abd27ef78b4fed9dc896457467" Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.139708905Z" level=info msg="cleanup sandbox network" Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.139823362Z" level=info msg="Successfully cleaned up network for pod 2cd51e2f6a1cbcdeb26d92c54070ab2a85b957f76038eb175b75e3f2ae639ed7" Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.139907014Z" level=info msg="cleanup sandbox network" Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.140440834Z" level=info msg="Got pod network &{Name:cni-sysctl-allowlist-ds-tpxqs Namespace:openshift-multus ID:1e1cece1b89c1fd30700275a80eef602cac31d5d3e2d8e7903561c25eafb4a33 UID:02e0075b-c660-4dc5-8cce-00fc4ef2980f NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.140741892Z" level=info msg="Deleting pod openshift-multus_cni-sysctl-allowlist-ds-tpxqs from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:30.211162 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xmq2g" event=&{ID:f364a949-09d7-4cfb-83ff-e532b822a557 Type:ContainerStarted Data:ed965039c2d980207be500acfdc76412b71aa85c7c7bd8c0df6174dd65016a5d} Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:30.211217 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xmq2g" event=&{ID:f364a949-09d7-4cfb-83ff-e532b822a557 Type:ContainerStarted Data:00fb6557a9939b9a2bc7c25ec8cc1770b954010cd506656807f39f57849767a8} Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:30.211449 2171 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:30.219677 2171 generic.go:296] "Generic (PLEG): container finished" podID=c0e47249-a5ff-460e-958d-53577387c154 containerID="7578d7bf6cd2d2172abb6b188e73d8fb8ff6b5b8946d9c2da442ed1cc948ab6f" exitCode=0 Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:30.219773 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerDied Data:7578d7bf6cd2d2172abb6b188e73d8fb8ff6b5b8946d9c2da442ed1cc948ab6f} Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.221274078Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:a8bcc8cf428e222c7e6677fe5f7f05e79da2aed1ded851cad696d83c94860b4c" id=1d3f799d-ba91-416c-a7fa-95a287bb2882 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.221994317Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:34fa51fa29c1ebd2a3ad620255bb70edf89b0a96348f08efac3dc2ca3200602d,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:a8bcc8cf428e222c7e6677fe5f7f05e79da2aed1ded851cad696d83c94860b4c],Size_:275275974,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=1d3f799d-ba91-416c-a7fa-95a287bb2882 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.231089186Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:a8bcc8cf428e222c7e6677fe5f7f05e79da2aed1ded851cad696d83c94860b4c" id=e74c1f42-942e-45fc-a543-e220614581da name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.231727978Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:34fa51fa29c1ebd2a3ad620255bb70edf89b0a96348f08efac3dc2ca3200602d,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:a8bcc8cf428e222c7e6677fe5f7f05e79da2aed1ded851cad696d83c94860b4c],Size_:275275974,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=e74c1f42-942e-45fc-a543-e220614581da name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.243865492Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/routeoverride-cni" id=155a7636-de8f-4ec9-947d-dac3f39ff69d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.243998296Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:30.247227 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hq594" event=&{ID:50152296-9489-4fa8-aa42-7858debe1c08 Type:ContainerStarted Data:8e2d55780ef7738a05fee323e5c71bd87390c87bd4b55a419d3d7362eb940075} Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:30.252156 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-8xrbh" event=&{ID:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748 Type:ContainerStarted Data:5ef6063b2be06818415c31d58e69b088971596d2dfdc9a42d50d5c9b5992d8e2} Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:30.252184 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-8xrbh" event=&{ID:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748 Type:ContainerStarted Data:16c0a5a6c5f7e9eecc5461c963ba5b26d7e13b2c8571603f6e05ef86ce3505e3} Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: 2023-01-05T09:16:30Z [verbose] Del: openshift-multus:cni-sysctl-allowlist-ds-tpxqs:unknownUID:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.289982141Z" level=info msg="Successfully cleaned up network for pod 1e1cece1b89c1fd30700275a80eef602cac31d5d3e2d8e7903561c25eafb4a33" Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.290070364Z" level=info msg="cleanup sandbox network" Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.290125406Z" level=info msg="Successfully cleaned up network for pod 0b984f430b21595a59efe2294b0b9fce6559367942c6743e20dbab9826dae909" Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.290183258Z" level=info msg="cleanup sandbox network" Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.290591246Z" level=info msg="Got pod network &{Name:ingress-canary-8xrbh Namespace:openshift-ingress-canary ID:3e1262384ce877333f952ac7ba12d68b4133181f637b84095f42300e1641af9e UID:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748 NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.290734027Z" level=info msg="Deleting pod openshift-ingress-canary_ingress-canary-8xrbh from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.430221390Z" level=info msg="Created container de8edae022873ac9a3e9319e361859b9ee302dea12c0b410170ede0731345f87: openshift-multus/network-metrics-daemon-hq594/network-metrics-daemon" id=5ed7d158-dc25-4cd1-a727-400cc79d752d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.433770682Z" level=info msg="Starting container: de8edae022873ac9a3e9319e361859b9ee302dea12c0b410170ede0731345f87" id=9cd413c2-4897-436f-a361-7c9d578e343d name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: 2023-01-05T09:16:30Z [verbose] Del: openshift-ingress-canary:ingress-canary-8xrbh:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.449264515Z" level=info msg="Started container" PID=4036 containerID=de8edae022873ac9a3e9319e361859b9ee302dea12c0b410170ede0731345f87 description=openshift-multus/network-metrics-daemon-hq594/network-metrics-daemon id=9cd413c2-4897-436f-a361-7c9d578e343d name=/runtime.v1.RuntimeService/StartContainer sandboxID=8e2d55780ef7738a05fee323e5c71bd87390c87bd4b55a419d3d7362eb940075 Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.451003335Z" level=info msg="Successfully cleaned up network for pod 3e1262384ce877333f952ac7ba12d68b4133181f637b84095f42300e1641af9e" Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.472454535Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=7ec1fe64-a5a0-4437-afe0-f979df49c54d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.472669296Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=7ec1fe64-a5a0-4437-afe0-f979df49c54d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.474722606Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=a933631e-e098-498e-97ce-88866ccd4dc1 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.474998399Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=a933631e-e098-498e-97ce-88866ccd4dc1 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.476421923Z" level=info msg="Creating container: openshift-multus/network-metrics-daemon-hq594/kube-rbac-proxy" id=e406244a-82f4-46fa-96aa-084fbcca5dc2 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.476542656Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:30.507745 2171 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-dns/dns-default-2pfzf] Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: 2023-01-05T09:16:30Z [verbose] Add: openshift-dns:dns-default-2pfzf:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/e6fe9a72-a6aa-41d5-86f7-be1f22acf516"}],"ips":[{"version":"4","interface":0,"address":"10.129.2.5/23","gateway":"10.129.2.1"}],"dns":{}} Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: I0105 09:16:30.481588 3575 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-dns", Name:"dns-default-2pfzf", UID:"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63", APIVersion:"v1", ResourceVersion:"79303", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.129.2.5/23] from openshift-sdn Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.509341481Z" level=info msg="Got pod network &{Name:dns-default-2pfzf Namespace:openshift-dns ID:63f26fe630a86618ccb6926337ae0f77635a06eec51dd1cb5b91c86eaefd519e UID:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 NetNS:/var/run/netns/e6fe9a72-a6aa-41d5-86f7-be1f22acf516 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.509692892Z" level=info msg="Checking pod openshift-dns_dns-default-2pfzf for CNI network multus-cni-network (type=multus)" Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.518060464Z" level=info msg="Ran pod sandbox 63f26fe630a86618ccb6926337ae0f77635a06eec51dd1cb5b91c86eaefd519e with infra container: openshift-dns/dns-default-2pfzf/POD" id=4a35ce7e-1e21-4eef-8938-665128689f0f name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.536440431Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f674b597694b13b7060c9ee611a403f1f87a98e05682c33f9a86140061ee9d9e" id=88174046-13cd-4f21-bb51-578530dc922c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.536867282Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:cdb247afcbd9968f1a59680dcc4de0364f7ddd8fc25dd3a12db59f05c8e026a0,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f674b597694b13b7060c9ee611a403f1f87a98e05682c33f9a86140061ee9d9e],Size_:404212434,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=88174046-13cd-4f21-bb51-578530dc922c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.538599292Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f674b597694b13b7060c9ee611a403f1f87a98e05682c33f9a86140061ee9d9e" id=71fb31ae-6172-4008-ae5a-d9023605d499 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.539827664Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:cdb247afcbd9968f1a59680dcc4de0364f7ddd8fc25dd3a12db59f05c8e026a0,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f674b597694b13b7060c9ee611a403f1f87a98e05682c33f9a86140061ee9d9e],Size_:404212434,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=71fb31ae-6172-4008-ae5a-d9023605d499 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.541457580Z" level=info msg="Creating container: openshift-dns/dns-default-2pfzf/dns" id=034cafd3-cb62-4113-a42e-89431c517c28 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.541564846Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.835550639Z" level=info msg="Created container dc9249bdde92d571096a926f7684713f098bdf7ed4b41fd5a676e2d05fb9ed88: openshift-multus/multus-additional-cni-plugins-5fgxc/routeoverride-cni" id=155a7636-de8f-4ec9-947d-dac3f39ff69d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.840877634Z" level=info msg="Starting container: dc9249bdde92d571096a926f7684713f098bdf7ed4b41fd5a676e2d05fb9ed88" id=b7ce1d70-cc3e-47f0-b921-7a919f208ee0 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.875431468Z" level=info msg="Started container" PID=4119 containerID=dc9249bdde92d571096a926f7684713f098bdf7ed4b41fd5a676e2d05fb9ed88 description=openshift-multus/multus-additional-cni-plugins-5fgxc/routeoverride-cni id=b7ce1d70-cc3e-47f0-b921-7a919f208ee0 name=/runtime.v1.RuntimeService/StartContainer sandboxID=01ea8cd190538ff3846bfa2fc9b101a796149e8b3c9ebd27850d03c9ae1f894d Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.926500810Z" level=info msg="Created container f40fcd06a4e10645ee4977d1ee4983cf57775ec396d60b9911807490ff4600f6: openshift-multus/network-metrics-daemon-hq594/kube-rbac-proxy" id=e406244a-82f4-46fa-96aa-084fbcca5dc2 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.927158054Z" level=info msg="Starting container: f40fcd06a4e10645ee4977d1ee4983cf57775ec396d60b9911807490ff4600f6" id=aeaf1846-672c-4394-a976-8087386bc438 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.940148124Z" level=info msg="Started container" PID=4148 containerID=f40fcd06a4e10645ee4977d1ee4983cf57775ec396d60b9911807490ff4600f6 description=openshift-multus/network-metrics-daemon-hq594/kube-rbac-proxy id=aeaf1846-672c-4394-a976-8087386bc438 name=/runtime.v1.RuntimeService/StartContainer sandboxID=8e2d55780ef7738a05fee323e5c71bd87390c87bd4b55a419d3d7362eb940075 Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.940734285Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_74d5af9c-b6bf-4922-8f4c-1cc489327e36\"" Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.972441695Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.972470805Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:30.972488555Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/route-override\"" Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:31.008607409Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:31.008644291Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:31.008667540Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_74d5af9c-b6bf-4922-8f4c-1cc489327e36\"" Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:31.017493415Z" level=info msg="Created container 4d6222fe563c720e5609103b6289c922e377aa7810fdc23632f5ecb0a4ae9973: openshift-dns/dns-default-2pfzf/dns" id=034cafd3-cb62-4113-a42e-89431c517c28 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:31.018350538Z" level=info msg="Starting container: 4d6222fe563c720e5609103b6289c922e377aa7810fdc23632f5ecb0a4ae9973" id=92f04f22-1e31-494d-b98c-231ed1654d1f name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:31.030854598Z" level=info msg="Started container" PID=4184 containerID=4d6222fe563c720e5609103b6289c922e377aa7810fdc23632f5ecb0a4ae9973 description=openshift-dns/dns-default-2pfzf/dns id=92f04f22-1e31-494d-b98c-231ed1654d1f name=/runtime.v1.RuntimeService/StartContainer sandboxID=63f26fe630a86618ccb6926337ae0f77635a06eec51dd1cb5b91c86eaefd519e Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:31.039865168Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=bfa06daa-e5bf-4753-9d0c-2b88884bb589 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:31.040380307Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=bfa06daa-e5bf-4753-9d0c-2b88884bb589 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:31.045163698Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=a434531c-4278-400a-a16d-d9c6e8eb5818 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:31.045499716Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=a434531c-4278-400a-a16d-d9c6e8eb5818 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:31.046619486Z" level=info msg="Creating container: openshift-dns/dns-default-2pfzf/kube-rbac-proxy" id=82ac4849-42db-48a0-b95d-872354607f7e name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:31.046796831Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:31.255196 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hq594" event=&{ID:50152296-9489-4fa8-aa42-7858debe1c08 Type:ContainerStarted Data:f40fcd06a4e10645ee4977d1ee4983cf57775ec396d60b9911807490ff4600f6} Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:31.256030 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hq594" event=&{ID:50152296-9489-4fa8-aa42-7858debe1c08 Type:ContainerStarted Data:de8edae022873ac9a3e9319e361859b9ee302dea12c0b410170ede0731345f87} Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:31.257271 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-2pfzf" event=&{ID:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 Type:ContainerStarted Data:4d6222fe563c720e5609103b6289c922e377aa7810fdc23632f5ecb0a4ae9973} Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:31.257449 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-2pfzf" event=&{ID:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 Type:ContainerStarted Data:63f26fe630a86618ccb6926337ae0f77635a06eec51dd1cb5b91c86eaefd519e} Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:31.262783 2171 generic.go:296] "Generic (PLEG): container finished" podID=c0e47249-a5ff-460e-958d-53577387c154 containerID="dc9249bdde92d571096a926f7684713f098bdf7ed4b41fd5a676e2d05fb9ed88" exitCode=0 Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:31.262923 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerDied Data:dc9249bdde92d571096a926f7684713f098bdf7ed4b41fd5a676e2d05fb9ed88} Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:31.264165581Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451" id=c9ee3319-0e2e-4606-90c6-0a9c8c4a4a86 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:31.264953732Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:56f4df3c126a569448a12d5112b7c20e99e219b2b2fe6a04c4ba743396fd7751,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451],Size_:445620510,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=c9ee3319-0e2e-4606-90c6-0a9c8c4a4a86 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:31.265732553Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451" id=6ad66435-344f-4082-95fb-725317392708 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:31.266044769Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:56f4df3c126a569448a12d5112b7c20e99e219b2b2fe6a04c4ba743396fd7751,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451],Size_:445620510,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=6ad66435-344f-4082-95fb-725317392708 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:31.268652185Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/whereabouts-cni-bincopy" id=5a2e10e4-9344-4c24-befc-9f9fcbcf20be name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:31.268942370Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:31.336814396Z" level=info msg="Created container 2f4396a38a0e99697d880e86d1714935e0fba0a6ee9dae69e5c548e56058a991: openshift-dns/dns-default-2pfzf/kube-rbac-proxy" id=82ac4849-42db-48a0-b95d-872354607f7e name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:31.340916590Z" level=info msg="Starting container: 2f4396a38a0e99697d880e86d1714935e0fba0a6ee9dae69e5c548e56058a991" id=cf179881-99f8-4ad4-950a-704841a3858c name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:31.378759634Z" level=info msg="Started container" PID=4274 containerID=2f4396a38a0e99697d880e86d1714935e0fba0a6ee9dae69e5c548e56058a991 description=openshift-dns/dns-default-2pfzf/kube-rbac-proxy id=cf179881-99f8-4ad4-950a-704841a3858c name=/runtime.v1.RuntimeService/StartContainer sandboxID=63f26fe630a86618ccb6926337ae0f77635a06eec51dd1cb5b91c86eaefd519e Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:31.545159441Z" level=info msg="Created container 0daa2075b1c32d916cacb66d9affcb5cdbfc3011fbc9855cc48d6576a9a80e99: openshift-multus/multus-additional-cni-plugins-5fgxc/whereabouts-cni-bincopy" id=5a2e10e4-9344-4c24-befc-9f9fcbcf20be name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:31.545682464Z" level=info msg="Starting container: 0daa2075b1c32d916cacb66d9affcb5cdbfc3011fbc9855cc48d6576a9a80e99" id=df7dd099-5aac-4f9e-8a93-61e8aeb5f3e4 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:31.558497679Z" level=info msg="Started container" PID=4322 containerID=0daa2075b1c32d916cacb66d9affcb5cdbfc3011fbc9855cc48d6576a9a80e99 description=openshift-multus/multus-additional-cni-plugins-5fgxc/whereabouts-cni-bincopy id=df7dd099-5aac-4f9e-8a93-61e8aeb5f3e4 name=/runtime.v1.RuntimeService/StartContainer sandboxID=01ea8cd190538ff3846bfa2fc9b101a796149e8b3c9ebd27850d03c9ae1f894d Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:31.573297896Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_35617c8e-4df6-4e16-abec-1400a6664672\"" Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:31.588511634Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:31.588561243Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:31.689001374Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/whereabouts\"" Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:31.709617698Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:31.709648592Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:16:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:31.709667939Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_35617c8e-4df6-4e16-abec-1400a6664672\"" Jan 05 09:16:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:32.267813 2171 generic.go:296] "Generic (PLEG): container finished" podID=c0e47249-a5ff-460e-958d-53577387c154 containerID="0daa2075b1c32d916cacb66d9affcb5cdbfc3011fbc9855cc48d6576a9a80e99" exitCode=0 Jan 05 09:16:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:32.268113 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerDied Data:0daa2075b1c32d916cacb66d9affcb5cdbfc3011fbc9855cc48d6576a9a80e99} Jan 05 09:16:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:32.268475250Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451" id=9fade86b-7b25-4364-9b8b-029a3f6ff76c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:32.268696699Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:56f4df3c126a569448a12d5112b7c20e99e219b2b2fe6a04c4ba743396fd7751,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451],Size_:445620510,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=9fade86b-7b25-4364-9b8b-029a3f6ff76c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:32.269475478Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451" id=b3b688a5-6b51-453d-a436-ed93e78591e5 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:32.269660972Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:56f4df3c126a569448a12d5112b7c20e99e219b2b2fe6a04c4ba743396fd7751,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451],Size_:445620510,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b3b688a5-6b51-453d-a436-ed93e78591e5 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:32.270710631Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/whereabouts-cni" id=4e46d7fe-2e2b-4687-af0e-2e98533b7a39 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:32.270822489Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:32.274169 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-2pfzf" event=&{ID:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 Type:ContainerStarted Data:2f4396a38a0e99697d880e86d1714935e0fba0a6ee9dae69e5c548e56058a991} Jan 05 09:16:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:32.437949975Z" level=info msg="Created container e64ea5f758693191c1782c7d7ec1a85373d64d6b7112e729bb7e44a6a023cf07: openshift-multus/multus-additional-cni-plugins-5fgxc/whereabouts-cni" id=4e46d7fe-2e2b-4687-af0e-2e98533b7a39 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:32.438376523Z" level=info msg="Starting container: e64ea5f758693191c1782c7d7ec1a85373d64d6b7112e729bb7e44a6a023cf07" id=1690a5bb-d907-42ff-a9b0-97ed38360879 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:32.444848704Z" level=info msg="Started container" PID=4442 containerID=e64ea5f758693191c1782c7d7ec1a85373d64d6b7112e729bb7e44a6a023cf07 description=openshift-multus/multus-additional-cni-plugins-5fgxc/whereabouts-cni id=1690a5bb-d907-42ff-a9b0-97ed38360879 name=/runtime.v1.RuntimeService/StartContainer sandboxID=01ea8cd190538ff3846bfa2fc9b101a796149e8b3c9ebd27850d03c9ae1f894d Jan 05 09:16:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:33.139270 2171 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeReady" Jan 05 09:16:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:33.139360 2171 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeSchedulable" Jan 05 09:16:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:33.183302 2171 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-2pfzf" Jan 05 09:16:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:33.277620 2171 generic.go:296] "Generic (PLEG): container finished" podID=c0e47249-a5ff-460e-958d-53577387c154 containerID="e64ea5f758693191c1782c7d7ec1a85373d64d6b7112e729bb7e44a6a023cf07" exitCode=0 Jan 05 09:16:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:33.277717 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerDied Data:e64ea5f758693191c1782c7d7ec1a85373d64d6b7112e729bb7e44a6a023cf07} Jan 05 09:16:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:33.279249720Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=069be91e-7bde-4c5b-ada5-20b480b59e81 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:33.279463162Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=069be91e-7bde-4c5b-ada5-20b480b59e81 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:33.280111728Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=0941080c-7257-4fa6-854e-70d2a28d5d80 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:33.280273545Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=0941080c-7257-4fa6-854e-70d2a28d5d80 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:33.281073073Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/kube-multus-additional-cni-plugins" id=660b010f-6332-4594-8556-f1fd16cb4ead name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:33.281180671Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:33.437038503Z" level=info msg="Created container 50791b1c54131e0dbc79dc92e6a9e2d86016118dabf05d21d38cb7dbe8398f4b: openshift-multus/multus-additional-cni-plugins-5fgxc/kube-multus-additional-cni-plugins" id=660b010f-6332-4594-8556-f1fd16cb4ead name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:33.437744197Z" level=info msg="Starting container: 50791b1c54131e0dbc79dc92e6a9e2d86016118dabf05d21d38cb7dbe8398f4b" id=8c87ff5d-825f-4c71-a049-220f8ae1f1fd name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:33.444818188Z" level=info msg="Started container" PID=4508 containerID=50791b1c54131e0dbc79dc92e6a9e2d86016118dabf05d21d38cb7dbe8398f4b description=openshift-multus/multus-additional-cni-plugins-5fgxc/kube-multus-additional-cni-plugins id=8c87ff5d-825f-4c71-a049-220f8ae1f1fd name=/runtime.v1.RuntimeService/StartContainer sandboxID=01ea8cd190538ff3846bfa2fc9b101a796149e8b3c9ebd27850d03c9ae1f894d Jan 05 09:16:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:34.281150 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerStarted Data:50791b1c54131e0dbc79dc92e6a9e2d86016118dabf05d21d38cb7dbe8398f4b} Jan 05 09:16:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:42.183870 2171 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-2pfzf" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.064819 2171 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-monitoring/kube-state-metrics-6798b548b9-xl88q] Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.064868 2171 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.068903 2171 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-monitoring/telemeter-client-5598c645c4-mftdv] Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.068940 2171 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.085456 2171 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh] Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.085503 2171 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.085794 2171 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256] Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.085817 2171 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.138039 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-state-metrics-tls\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.138104 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-telemeter-client-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-secret-telemeter-client-kube-rbac-proxy-config\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.138149 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r64p\" (UniqueName: \"kubernetes.io/projected/fc45e847-8f55-4c48-8c48-4bfee2419dca-kube-api-access-4r64p\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.138194 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-metrics-client-ca\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.138233 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/a40bf20d-67a4-45e7-b701-ca5a75376c1f-audit-log\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.138275 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-config\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.138323 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a40bf20d-67a4-45e7-b701-ca5a75376c1f-tmpfs\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.138365 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-metrics-client-ca\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.138462 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-telemeter-client\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-secret-telemeter-client\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.138506 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-openshift-state-metrics-tls\") pod \"openshift-state-metrics-6f5dcdf79b-wq256\" (UID: \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\") " pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.138552 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-6f5dcdf79b-wq256\" (UID: \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\") " pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.138598 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6g5r\" (UniqueName: \"kubernetes.io/projected/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-kube-api-access-h6g5r\") pod \"openshift-state-metrics-6f5dcdf79b-wq256\" (UID: \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\") " pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.138641 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemeter-client-tls\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-telemeter-client-tls\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.138688 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfs6g\" (UniqueName: \"kubernetes.io/projected/a40bf20d-67a4-45e7-b701-ca5a75376c1f-kube-api-access-gfs6g\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.138737 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.138803 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemeter-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-telemeter-trusted-ca-bundle\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.138847 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-volume-directive-shadow\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.138889 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-serving-certs-ca-bundle\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.138933 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blpwr\" (UniqueName: \"kubernetes.io/projected/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-api-access-blpwr\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.138981 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-adapter-prometheus-config\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-prometheus-adapter-prometheus-config\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.139030 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls\" (UniqueName: \"kubernetes.io/secret/a40bf20d-67a4-45e7-b701-ca5a75376c1f-tls\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.139076 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-serving-certs-ca-bundle\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.139124 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-adapter-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-prometheus-adapter-audit-profiles\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.223098 2171 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn] Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.223139 2171 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.239371 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-gfs6g\" (UniqueName: \"kubernetes.io/projected/a40bf20d-67a4-45e7-b701-ca5a75376c1f-kube-api-access-gfs6g\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.239463 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.239500 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-volume-directive-shadow\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.239533 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"telemeter-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-telemeter-trusted-ca-bundle\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.239566 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-blpwr\" (UniqueName: \"kubernetes.io/projected/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-api-access-blpwr\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.239603 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-serving-certs-ca-bundle\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.239640 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"prometheus-adapter-prometheus-config\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-prometheus-adapter-prometheus-config\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.239677 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tls\" (UniqueName: \"kubernetes.io/secret/a40bf20d-67a4-45e7-b701-ca5a75376c1f-tls\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.239713 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-serving-certs-ca-bundle\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.239775 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"prometheus-adapter-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-prometheus-adapter-audit-profiles\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.239813 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-telemeter-client-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-secret-telemeter-client-kube-rbac-proxy-config\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.239847 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-4r64p\" (UniqueName: \"kubernetes.io/projected/fc45e847-8f55-4c48-8c48-4bfee2419dca-kube-api-access-4r64p\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.239886 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-state-metrics-tls\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.239918 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-config\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.239954 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-metrics-client-ca\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.239988 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/a40bf20d-67a4-45e7-b701-ca5a75376c1f-audit-log\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.240025 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a40bf20d-67a4-45e7-b701-ca5a75376c1f-tmpfs\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.240062 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-metrics-client-ca\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.240106 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-openshift-state-metrics-tls\") pod \"openshift-state-metrics-6f5dcdf79b-wq256\" (UID: \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\") " pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.240144 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-6f5dcdf79b-wq256\" (UID: \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\") " pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.240177 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-telemeter-client\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-secret-telemeter-client\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.240212 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-h6g5r\" (UniqueName: \"kubernetes.io/projected/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-kube-api-access-h6g5r\") pod \"openshift-state-metrics-6f5dcdf79b-wq256\" (UID: \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\") " pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.240247 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"telemeter-client-tls\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-telemeter-client-tls\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.242023 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"prometheus-adapter-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-prometheus-adapter-audit-profiles\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.246856 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"telemeter-client-tls\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-telemeter-client-tls\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.248135 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-volume-directive-shadow\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.258490 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-serving-certs-ca-bundle\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.259068 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"prometheus-adapter-prometheus-config\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-prometheus-adapter-prometheus-config\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.262042 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"telemeter-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-telemeter-trusted-ca-bundle\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.263605 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-serving-certs-ca-bundle\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.263895 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a40bf20d-67a4-45e7-b701-ca5a75376c1f-tmpfs\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.266521 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-state-metrics-tls\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.267098 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-config\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.267661 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-metrics-client-ca\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.282126 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-metrics-client-ca\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.282324 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/a40bf20d-67a4-45e7-b701-ca5a75376c1f-audit-log\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.283447 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-telemeter-client-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-secret-telemeter-client-kube-rbac-proxy-config\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.297913 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-openshift-state-metrics-tls\") pod \"openshift-state-metrics-6f5dcdf79b-wq256\" (UID: \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\") " pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.298499 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfs6g\" (UniqueName: \"kubernetes.io/projected/a40bf20d-67a4-45e7-b701-ca5a75376c1f-kube-api-access-gfs6g\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.299055 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tls\" (UniqueName: \"kubernetes.io/secret/a40bf20d-67a4-45e7-b701-ca5a75376c1f-tls\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.299505 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.299922 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-telemeter-client\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-secret-telemeter-client\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.302924 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-6f5dcdf79b-wq256\" (UID: \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\") " pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.305941 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-blpwr\" (UniqueName: \"kubernetes.io/projected/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-api-access-blpwr\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.307702 2171 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-ingress/router-default-5cf6bbdff9-tnffs] Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.307743 2171 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.340980 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccjjw\" (UniqueName: \"kubernetes.io/projected/c3b10ff4-a243-4a4e-b27d-f613e258666d-kube-api-access-ccjjw\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.341025 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-default-certificate\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.341064 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-stats-auth\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.341135 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-metrics-certs\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.341169 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3b10ff4-a243-4a4e-b27d-f613e258666d-service-ca-bundle\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.341316 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/88745ab7-efe7-42c0-a2e8-75591518333d-tls-certificates\") pod \"prometheus-operator-admission-webhook-598468bb49-94dsn\" (UID: \"88745ab7-efe7-42c0-a2e8-75591518333d\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.344304 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r64p\" (UniqueName: \"kubernetes.io/projected/fc45e847-8f55-4c48-8c48-4bfee2419dca-kube-api-access-4r64p\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.347169 2171 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l] Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.347205 2171 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.364249 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6g5r\" (UniqueName: \"kubernetes.io/projected/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-kube-api-access-h6g5r\") pod \"openshift-state-metrics-6f5dcdf79b-wq256\" (UID: \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\") " pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.442289 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.442335 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.442375 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.442436 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-oauth-cookie\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-oauth-cookie\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.442469 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/7c63c9be-d1ad-44b6-8b33-09af3ee314af-metrics-client-ca\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.442512 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/88745ab7-efe7-42c0-a2e8-75591518333d-tls-certificates\") pod \"prometheus-operator-admission-webhook-598468bb49-94dsn\" (UID: \"88745ab7-efe7-42c0-a2e8-75591518333d\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.442550 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-stats-auth\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.442590 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3b10ff4-a243-4a4e-b27d-f613e258666d-service-ca-bundle\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.442624 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvz7l\" (UniqueName: \"kubernetes.io/projected/7c63c9be-d1ad-44b6-8b33-09af3ee314af-kube-api-access-pvz7l\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.442662 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-querier-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c63c9be-d1ad-44b6-8b33-09af3ee314af-thanos-querier-trusted-ca-bundle\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.442710 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-tls\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.442742 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-grpc-tls\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.442791 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-ccjjw\" (UniqueName: \"kubernetes.io/projected/c3b10ff4-a243-4a4e-b27d-f613e258666d-kube-api-access-ccjjw\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.442827 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-default-certificate\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.442862 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-metrics-certs\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.444160 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3b10ff4-a243-4a4e-b27d-f613e258666d-service-ca-bundle\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.445328 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-metrics-certs\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.447535 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-stats-auth\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.447535 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/88745ab7-efe7-42c0-a2e8-75591518333d-tls-certificates\") pod \"prometheus-operator-admission-webhook-598468bb49-94dsn\" (UID: \"88745ab7-efe7-42c0-a2e8-75591518333d\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.449150 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-default-certificate\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.450949 2171 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:50.451354587Z" level=info msg="Running pod sandbox: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/POD" id=651563b1-ecd9-4037-a40a-5374a5f401ac name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:50.451431874Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.463450 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccjjw\" (UniqueName: \"kubernetes.io/projected/c3b10ff4-a243-4a4e-b27d-f613e258666d-kube-api-access-ccjjw\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:50.468910315Z" level=info msg="Got pod network &{Name:kube-state-metrics-6798b548b9-xl88q Namespace:openshift-monitoring ID:bb417fcfe195906c7b26f0d9c20687f555b79c545c37fe3e4cebd9d60e23d18c UID:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e NetNS:/var/run/netns/4475bb08-045b-4d37-bc48-4d8025ed2f99 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:50.468935863Z" level=info msg="Adding pod openshift-monitoring_kube-state-metrics-6798b548b9-xl88q to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.476727 2171 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:50.478086185Z" level=info msg="Running pod sandbox: openshift-monitoring/telemeter-client-5598c645c4-mftdv/POD" id=3e0224d0-1c16-43fd-b2b2-8bb2b8074726 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:50.478119323Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:50.494333697Z" level=info msg="Got pod network &{Name:telemeter-client-5598c645c4-mftdv Namespace:openshift-monitoring ID:0a5caa2be3488b9c8c36d43ddc5d5a1784a28176b83143a0b5296da42812377f UID:fc45e847-8f55-4c48-8c48-4bfee2419dca NetNS:/var/run/netns/21c06552-1790-4829-956f-9ac9d4e5ebe7 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:50.494366157Z" level=info msg="Adding pod openshift-monitoring_telemeter-client-5598c645c4-mftdv to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.505012 2171 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:50.505618273Z" level=info msg="Running pod sandbox: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/POD" id=922e2bf1-84b5-4b15-afac-34f0508ae49d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:50.505663869Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.521211 2171 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:50.521765965Z" level=info msg="Running pod sandbox: openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh/POD" id=fd87ff3e-de92-4210-a501-d832c297737c name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:50.521812199Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.544892 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-tls\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.544945 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-grpc-tls\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.544990 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.545024 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.545064 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.545103 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-oauth-cookie\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-oauth-cookie\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.545136 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/7c63c9be-d1ad-44b6-8b33-09af3ee314af-metrics-client-ca\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.545183 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-pvz7l\" (UniqueName: \"kubernetes.io/projected/7c63c9be-d1ad-44b6-8b33-09af3ee314af-kube-api-access-pvz7l\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.545219 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"thanos-querier-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c63c9be-d1ad-44b6-8b33-09af3ee314af-thanos-querier-trusted-ca-bundle\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.549800 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.550098 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.550361 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"thanos-querier-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c63c9be-d1ad-44b6-8b33-09af3ee314af-thanos-querier-trusted-ca-bundle\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.550931 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/7c63c9be-d1ad-44b6-8b33-09af3ee314af-metrics-client-ca\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.552600 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-oauth-cookie\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-oauth-cookie\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.578973 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-tls\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.580321 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-grpc-tls\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.580655 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.582641 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvz7l\" (UniqueName: \"kubernetes.io/projected/7c63c9be-d1ad-44b6-8b33-09af3ee314af-kube-api-access-pvz7l\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:50.597004482Z" level=info msg="Got pod network &{Name:openshift-state-metrics-6f5dcdf79b-wq256 Namespace:openshift-monitoring ID:1ffc39284205644ee2d20c66b8df819acc6bda29c81ecc932ae2237e02a208e7 UID:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 NetNS:/var/run/netns/6fa2a531-712f-4786-92e8-d0a01014b363 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:50.597045590Z" level=info msg="Adding pod openshift-monitoring_openshift-state-metrics-6f5dcdf79b-wq256 to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.646610 2171 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:50.647271521Z" level=info msg="Running pod sandbox: openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn/POD" id=7ceb7111-bb4a-4639-9f6d-a48ad3bf9d69 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:50.647326258Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.699214 2171 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:50.701001624Z" level=info msg="Running pod sandbox: openshift-ingress/router-default-5cf6bbdff9-tnffs/POD" id=6ddd2184-f7a9-49cc-8543-f9ddad731326 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:50.701055587Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:50.709790 2171 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:50.710555675Z" level=info msg="Running pod sandbox: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/POD" id=a1c28dad-bec7-4338-a18c-33bdf93f24a1 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:50.710597001Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:50.720834416Z" level=info msg="Got pod network &{Name:prometheus-adapter-5894cc86c7-qhmbh Namespace:openshift-monitoring ID:bf8032a2230aa8777b5ec76b9ff9222eb0d8d4700ec6b91316ee8cd870968d80 UID:a40bf20d-67a4-45e7-b701-ca5a75376c1f NetNS:/var/run/netns/a033356f-4162-40fb-ac68-ff35ebe0d82e Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:50.720867467Z" level=info msg="Adding pod openshift-monitoring_prometheus-adapter-5894cc86c7-qhmbh to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:50.770826172Z" level=info msg="Got pod network &{Name:prometheus-operator-admission-webhook-598468bb49-94dsn Namespace:openshift-monitoring ID:14626e8a3347256c5a502841426c6a0bfa7993e08a5404798831b7d49a3dddf7 UID:88745ab7-efe7-42c0-a2e8-75591518333d NetNS:/var/run/netns/86a5c5e8-22e7-4e2d-a573-d2c4609eca57 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:50.770866256Z" level=info msg="Adding pod openshift-monitoring_prometheus-operator-admission-webhook-598468bb49-94dsn to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:50.803794261Z" level=info msg="Got pod network &{Name:thanos-querier-6f4f5fb7cf-rdm9l Namespace:openshift-monitoring ID:f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3 UID:7c63c9be-d1ad-44b6-8b33-09af3ee314af NetNS:/var/run/netns/b93696a6-c824-4b48-ae0e-2e8a183d9b47 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:50.803834467Z" level=info msg="Adding pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:50.832324985Z" level=info msg="Got pod network &{Name:router-default-5cf6bbdff9-tnffs Namespace:openshift-ingress ID:d768276def4b37ab70feeee6455eb37f4fe6716bad6df283af0049c5724973e7 UID:c3b10ff4-a243-4a4e-b27d-f613e258666d NetNS:/var/run/netns/0f6c1c50-ff33-4781-9e31-36c023f9aee3 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:16:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:50.832550158Z" level=info msg="Adding pod openshift-ingress_router-default-5cf6bbdff9-tnffs to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:16:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:52.958098 2171 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/telemeter-client-5598c645c4-mftdv] Jan 05 09:16:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:52.962577 2171 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256] Jan 05 09:16:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:52.962616 2171 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/kube-state-metrics-6798b548b9-xl88q] Jan 05 09:16:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:52.962635 2171 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l] Jan 05 09:16:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: 2023-01-05T09:16:51Z [verbose] Add: openshift-monitoring:telemeter-client-5598c645c4-mftdv:fc45e847-8f55-4c48-8c48-4bfee2419dca:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/21c06552-1790-4829-956f-9ac9d4e5ebe7"}],"ips":[{"version":"4","interface":0,"address":"10.129.2.7/23","gateway":"10.129.2.1"}],"dns":{}} Jan 05 09:16:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: I0105 09:16:51.903071 4858 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"telemeter-client-5598c645c4-mftdv", UID:"fc45e847-8f55-4c48-8c48-4bfee2419dca", APIVersion:"v1", ResourceVersion:"79809", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.129.2.7/23] from openshift-sdn Jan 05 09:16:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:52.970016991Z" level=info msg="Got pod network &{Name:telemeter-client-5598c645c4-mftdv Namespace:openshift-monitoring ID:0a5caa2be3488b9c8c36d43ddc5d5a1784a28176b83143a0b5296da42812377f UID:fc45e847-8f55-4c48-8c48-4bfee2419dca NetNS:/var/run/netns/21c06552-1790-4829-956f-9ac9d4e5ebe7 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:16:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:52.970170998Z" level=info msg="Checking pod openshift-monitoring_telemeter-client-5598c645c4-mftdv for CNI network multus-cni-network (type=multus)" Jan 05 09:16:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: 2023-01-05T09:16:51Z [verbose] Add: openshift-monitoring:openshift-state-metrics-6f5dcdf79b-wq256:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/6fa2a531-712f-4786-92e8-d0a01014b363"}],"ips":[{"version":"4","interface":0,"address":"10.129.2.8/23","gateway":"10.129.2.1"}],"dns":{}} Jan 05 09:16:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: I0105 09:16:51.982087 4902 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"openshift-state-metrics-6f5dcdf79b-wq256", UID:"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6", APIVersion:"v1", ResourceVersion:"79822", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.129.2.8/23] from openshift-sdn Jan 05 09:16:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:52.971017063Z" level=info msg="Got pod network &{Name:openshift-state-metrics-6f5dcdf79b-wq256 Namespace:openshift-monitoring ID:1ffc39284205644ee2d20c66b8df819acc6bda29c81ecc932ae2237e02a208e7 UID:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 NetNS:/var/run/netns/6fa2a531-712f-4786-92e8-d0a01014b363 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:16:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:52.971141734Z" level=info msg="Checking pod openshift-monitoring_openshift-state-metrics-6f5dcdf79b-wq256 for CNI network multus-cni-network (type=multus)" Jan 05 09:16:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: 2023-01-05T09:16:52Z [verbose] Add: openshift-monitoring:thanos-querier-6f4f5fb7cf-rdm9l:7c63c9be-d1ad-44b6-8b33-09af3ee314af:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/b93696a6-c824-4b48-ae0e-2e8a183d9b47"}],"ips":[{"version":"4","interface":0,"address":"10.129.2.11/23","gateway":"10.129.2.1"}],"dns":{}} Jan 05 09:16:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: I0105 09:16:52.312457 4968 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"thanos-querier-6f4f5fb7cf-rdm9l", UID:"7c63c9be-d1ad-44b6-8b33-09af3ee314af", APIVersion:"v1", ResourceVersion:"79866", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.129.2.11/23] from openshift-sdn Jan 05 09:16:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:52.975236 2171 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh] Jan 05 09:16:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:52.979202 2171 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-ingress/router-default-5cf6bbdff9-tnffs] Jan 05 09:16:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:52.972330352Z" level=info msg="Got pod network &{Name:thanos-querier-6f4f5fb7cf-rdm9l Namespace:openshift-monitoring ID:f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3 UID:7c63c9be-d1ad-44b6-8b33-09af3ee314af NetNS:/var/run/netns/b93696a6-c824-4b48-ae0e-2e8a183d9b47 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:16:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:52.972475018Z" level=info msg="Checking pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l for CNI network multus-cni-network (type=multus)" Jan 05 09:16:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: 2023-01-05T09:16:51Z [verbose] Add: openshift-monitoring:kube-state-metrics-6798b548b9-xl88q:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/4475bb08-045b-4d37-bc48-4d8025ed2f99"}],"ips":[{"version":"4","interface":0,"address":"10.129.2.6/23","gateway":"10.129.2.1"}],"dns":{}} Jan 05 09:16:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: I0105 09:16:51.993309 4852 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"kube-state-metrics-6798b548b9-xl88q", UID:"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e", APIVersion:"v1", ResourceVersion:"79801", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.129.2.6/23] from openshift-sdn Jan 05 09:16:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:52.973923461Z" level=info msg="Got pod network &{Name:kube-state-metrics-6798b548b9-xl88q Namespace:openshift-monitoring ID:bb417fcfe195906c7b26f0d9c20687f555b79c545c37fe3e4cebd9d60e23d18c UID:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e NetNS:/var/run/netns/4475bb08-045b-4d37-bc48-4d8025ed2f99 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:16:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:52.974056231Z" level=info msg="Checking pod openshift-monitoring_kube-state-metrics-6798b548b9-xl88q for CNI network multus-cni-network (type=multus)" Jan 05 09:16:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:52.984263817Z" level=info msg="Ran pod sandbox bb417fcfe195906c7b26f0d9c20687f555b79c545c37fe3e4cebd9d60e23d18c with infra container: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/POD" id=651563b1-ecd9-4037-a40a-5374a5f401ac name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:52.986568027Z" level=info msg="Ran pod sandbox 0a5caa2be3488b9c8c36d43ddc5d5a1784a28176b83143a0b5296da42812377f with infra container: openshift-monitoring/telemeter-client-5598c645c4-mftdv/POD" id=3e0224d0-1c16-43fd-b2b2-8bb2b8074726 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:52.988914813Z" level=info msg="Ran pod sandbox f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3 with infra container: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/POD" id=a1c28dad-bec7-4338-a18c-33bdf93f24a1 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:52.991371010Z" level=info msg="Ran pod sandbox 1ffc39284205644ee2d20c66b8df819acc6bda29c81ecc932ae2237e02a208e7 with infra container: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/POD" id=922e2bf1-84b5-4b15-afac-34f0508ae49d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: 2023-01-05T09:16:52Z [verbose] Add: openshift-monitoring:prometheus-adapter-5894cc86c7-qhmbh:a40bf20d-67a4-45e7-b701-ca5a75376c1f:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/a033356f-4162-40fb-ac68-ff35ebe0d82e"}],"ips":[{"version":"4","interface":0,"address":"10.129.2.9/23","gateway":"10.129.2.1"}],"dns":{}} Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: I0105 09:16:52.162647 4939 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"prometheus-adapter-5894cc86c7-qhmbh", UID:"a40bf20d-67a4-45e7-b701-ca5a75376c1f", APIVersion:"v1", ResourceVersion:"79827", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.129.2.9/23] from openshift-sdn Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:52.999324874Z" level=info msg="Got pod network &{Name:prometheus-adapter-5894cc86c7-qhmbh Namespace:openshift-monitoring ID:bf8032a2230aa8777b5ec76b9ff9222eb0d8d4700ec6b91316ee8cd870968d80 UID:a40bf20d-67a4-45e7-b701-ca5a75376c1f NetNS:/var/run/netns/a033356f-4162-40fb-ac68-ff35ebe0d82e Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:52.999511962Z" level=info msg="Checking pod openshift-monitoring_prometheus-adapter-5894cc86c7-qhmbh for CNI network multus-cni-network (type=multus)" Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.001459512Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:85f671344a3bc1584db4caed39939b6dd55644a5bd806e118527f31dac91252e" id=b360d494-65ec-414e-84bf-b040447f4443 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.001680156Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:85f671344a3bc1584db4caed39939b6dd55644a5bd806e118527f31dac91252e not found" id=b360d494-65ec-414e-84bf-b040447f4443 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.001820098Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc" id=b09a178b-9f41-427d-b146-962e462c1563 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.002000598Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:08ff7a43d9185529520a29e114f0c28b070000b464a44b347240015f3b466488,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc],Size_:368278089,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b09a178b-9f41-427d-b146-962e462c1563 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: 2023-01-05T09:16:52Z [verbose] Add: openshift-ingress:router-default-5cf6bbdff9-tnffs:c3b10ff4-a243-4a4e-b27d-f613e258666d:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/0f6c1c50-ff33-4781-9e31-36c023f9aee3"}],"ips":[{"version":"4","interface":0,"address":"10.129.2.12/23","gateway":"10.129.2.1"}],"dns":{}} Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: I0105 09:16:52.670890 4978 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-ingress", Name:"router-default-5cf6bbdff9-tnffs", UID:"c3b10ff4-a243-4a4e-b27d-f613e258666d", APIVersion:"v1", ResourceVersion:"79863", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.129.2.12/23] from openshift-sdn Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.003278543Z" level=info msg="Got pod network &{Name:router-default-5cf6bbdff9-tnffs Namespace:openshift-ingress ID:d768276def4b37ab70feeee6455eb37f4fe6716bad6df283af0049c5724973e7 UID:c3b10ff4-a243-4a4e-b27d-f613e258666d NetNS:/var/run/netns/0f6c1c50-ff33-4781-9e31-36c023f9aee3 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.003440018Z" level=info msg="Checking pod openshift-ingress_router-default-5cf6bbdff9-tnffs for CNI network multus-cni-network (type=multus)" Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.008037608Z" level=info msg="Ran pod sandbox bf8032a2230aa8777b5ec76b9ff9222eb0d8d4700ec6b91316ee8cd870968d80 with infra container: openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh/POD" id=fd87ff3e-de92-4210-a501-d832c297737c name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.009329856Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc" id=18b0af4e-d5fb-4806-883d-f53f4f4a1a1b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.009554012Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:08ff7a43d9185529520a29e114f0c28b070000b464a44b347240015f3b466488,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc],Size_:368278089,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=18b0af4e-d5fb-4806-883d-f53f4f4a1a1b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.009637665Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=3c5ce3fc-2177-4289-be41-27b67163a4a0 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.009843703Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=3c5ce3fc-2177-4289-be41-27b67163a4a0 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.009900244Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e5521a3d1f64483e63ffe5bc23d9f882781cd4d152ac6de72013eaefe6dfcf89" id=9e6a174f-bd48-4fa1-98a9-826396b47397 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.010063531Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e5521a3d1f64483e63ffe5bc23d9f882781cd4d152ac6de72013eaefe6dfcf89 not found" id=9e6a174f-bd48-4fa1-98a9-826396b47397 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:53.011437 2171 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.016358094Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=97a8c20f-55a0-4a0c-a5a7-41b353aac949 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.016585044Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=97a8c20f-55a0-4a0c-a5a7-41b353aac949 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.016799163Z" level=info msg="Creating container: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/thanos-query" id=92197d1a-4e43-4f25-b27e-0b511cb5b7ca name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.016934715Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.020813642Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e5521a3d1f64483e63ffe5bc23d9f882781cd4d152ac6de72013eaefe6dfcf89" id=0f68ee03-c88b-45c9-8459-2e255fce5cd2 name=/runtime.v1.ImageService/PullImage Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.020990282Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:85f671344a3bc1584db4caed39939b6dd55644a5bd806e118527f31dac91252e" id=0c449ce7-8276-4cfa-8fef-4be300f09bba name=/runtime.v1.ImageService/PullImage Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.021033900Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9696f8f231d49aadf63ab980ec5c17c4fd87011930daecd30cc9e4ee102f1aeb" id=65fe8c59-ad2e-470d-a6fb-47712fb879f1 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 09:16:53.023682 2171 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3b10ff4_a243_4a4e_b27d_f613e258666d.slice/crio-d768276def4b37ab70feeee6455eb37f4fe6716bad6df283af0049c5724973e7.scope WatchSource:0}: Error finding container d768276def4b37ab70feeee6455eb37f4fe6716bad6df283af0049c5724973e7: Status 404 returned error can't find the container with id d768276def4b37ab70feeee6455eb37f4fe6716bad6df283af0049c5724973e7 Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.023729605Z" level=info msg="Creating container: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/kube-rbac-proxy-main" id=cb6dee0f-88b1-449a-996c-29fd0c1dc385 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.023868804Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.025516438Z" level=info msg="Ran pod sandbox d768276def4b37ab70feeee6455eb37f4fe6716bad6df283af0049c5724973e7 with infra container: openshift-ingress/router-default-5cf6bbdff9-tnffs/POD" id=6ddd2184-f7a9-49cc-8543-f9ddad731326 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.026527353Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:bfd14188641a2d34e5b2b2ed3336b3b440155533f3fd7c6de776300dc3ff9e3b" id=5946e877-2aa9-43d1-b383-2c09d56b781c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.033900903Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:ad00ebc82f5cc671785b8401477b42db226d1c335ea60a4bb39bf28a961de840,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:bfd14188641a2d34e5b2b2ed3336b3b440155533f3fd7c6de776300dc3ff9e3b],Size_:411774965,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=5946e877-2aa9-43d1-b383-2c09d56b781c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.034494854Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:ea0b364e3a8a4e4b3f4ab3288e42c5be8021187d32791029f9cb9a500c8da100,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9696f8f231d49aadf63ab980ec5c17c4fd87011930daecd30cc9e4ee102f1aeb],Size_:402351209,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=65fe8c59-ad2e-470d-a6fb-47712fb879f1 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.037337965Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:85f671344a3bc1584db4caed39939b6dd55644a5bd806e118527f31dac91252e\"" Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.037908776Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:bfd14188641a2d34e5b2b2ed3336b3b440155533f3fd7c6de776300dc3ff9e3b" id=b04e63bd-d546-435b-b5a3-a458eb7bf556 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.037972147Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e5521a3d1f64483e63ffe5bc23d9f882781cd4d152ac6de72013eaefe6dfcf89\"" Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.038076735Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:ad00ebc82f5cc671785b8401477b42db226d1c335ea60a4bb39bf28a961de840,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:bfd14188641a2d34e5b2b2ed3336b3b440155533f3fd7c6de776300dc3ff9e3b],Size_:411774965,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b04e63bd-d546-435b-b5a3-a458eb7bf556 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.038211762Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9696f8f231d49aadf63ab980ec5c17c4fd87011930daecd30cc9e4ee102f1aeb" id=49232d44-79b6-4c49-af85-de03c0fd11aa name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.038455183Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:ea0b364e3a8a4e4b3f4ab3288e42c5be8021187d32791029f9cb9a500c8da100,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9696f8f231d49aadf63ab980ec5c17c4fd87011930daecd30cc9e4ee102f1aeb],Size_:402351209,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=49232d44-79b6-4c49-af85-de03c0fd11aa name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.044019214Z" level=info msg="Creating container: openshift-ingress/router-default-5cf6bbdff9-tnffs/router" id=0fe92ec6-94cf-4a75-9e1e-9c407ee741c0 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.044140728Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.046302745Z" level=info msg="Creating container: openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh/prometheus-adapter" id=79cd50ca-13ea-4c82-acce-19ba3f0f4acc name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.046474812Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:53.054428 2171 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn] Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: 2023-01-05T09:16:52Z [verbose] Add: openshift-monitoring:prometheus-operator-admission-webhook-598468bb49-94dsn:88745ab7-efe7-42c0-a2e8-75591518333d:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/86a5c5e8-22e7-4e2d-a573-d2c4609eca57"}],"ips":[{"version":"4","interface":0,"address":"10.129.2.10/23","gateway":"10.129.2.1"}],"dns":{}} Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: I0105 09:16:52.973431 4957 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"prometheus-operator-admission-webhook-598468bb49-94dsn", UID:"88745ab7-efe7-42c0-a2e8-75591518333d", APIVersion:"v1", ResourceVersion:"79847", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.129.2.10/23] from openshift-sdn Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.071088386Z" level=info msg="Got pod network &{Name:prometheus-operator-admission-webhook-598468bb49-94dsn Namespace:openshift-monitoring ID:14626e8a3347256c5a502841426c6a0bfa7993e08a5404798831b7d49a3dddf7 UID:88745ab7-efe7-42c0-a2e8-75591518333d NetNS:/var/run/netns/86a5c5e8-22e7-4e2d-a573-d2c4609eca57 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.071211517Z" level=info msg="Checking pod openshift-monitoring_prometheus-operator-admission-webhook-598468bb49-94dsn for CNI network multus-cni-network (type=multus)" Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 09:16:53.092459 2171 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88745ab7_efe7_42c0_a2e8_75591518333d.slice/crio-14626e8a3347256c5a502841426c6a0bfa7993e08a5404798831b7d49a3dddf7.scope WatchSource:0}: Error finding container 14626e8a3347256c5a502841426c6a0bfa7993e08a5404798831b7d49a3dddf7: Status 404 returned error can't find the container with id 14626e8a3347256c5a502841426c6a0bfa7993e08a5404798831b7d49a3dddf7 Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.097873593Z" level=info msg="Ran pod sandbox 14626e8a3347256c5a502841426c6a0bfa7993e08a5404798831b7d49a3dddf7 with infra container: openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn/POD" id=7ceb7111-bb4a-4639-9f6d-a48ad3bf9d69 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.100361755Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:dbf46164dcc3b2b6ef5df5d043ddc4fd859fd2c73c4862808424b3df6a1e27e7" id=66c70a67-f093-4f7b-beb5-fabb6bf4d112 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.212502354Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:20e9403b34a0d9d25151a4ddbee3f69b99894fd8e058e6cb7310fe95080ecb03,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:dbf46164dcc3b2b6ef5df5d043ddc4fd859fd2c73c4862808424b3df6a1e27e7],Size_:374119253,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=66c70a67-f093-4f7b-beb5-fabb6bf4d112 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.231614669Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:dbf46164dcc3b2b6ef5df5d043ddc4fd859fd2c73c4862808424b3df6a1e27e7" id=977cf1e9-314f-4e54-ac9e-8ab8b01bfac3 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.231842274Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:20e9403b34a0d9d25151a4ddbee3f69b99894fd8e058e6cb7310fe95080ecb03,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:dbf46164dcc3b2b6ef5df5d043ddc4fd859fd2c73c4862808424b3df6a1e27e7],Size_:374119253,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=977cf1e9-314f-4e54-ac9e-8ab8b01bfac3 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.232676818Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:85f671344a3bc1584db4caed39939b6dd55644a5bd806e118527f31dac91252e\"" Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.233339512Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e5521a3d1f64483e63ffe5bc23d9f882781cd4d152ac6de72013eaefe6dfcf89\"" Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.234660218Z" level=info msg="Creating container: openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn/prometheus-operator-admission-webhook" id=20443b6b-06aa-47f6-b793-437c962a4a6d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.235065644Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:53.372712 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" event=&{ID:fc45e847-8f55-4c48-8c48-4bfee2419dca Type:ContainerStarted Data:0a5caa2be3488b9c8c36d43ddc5d5a1784a28176b83143a0b5296da42812377f} Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:53.373443 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" event=&{ID:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e Type:ContainerStarted Data:bb417fcfe195906c7b26f0d9c20687f555b79c545c37fe3e4cebd9d60e23d18c} Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:53.374111 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" event=&{ID:c3b10ff4-a243-4a4e-b27d-f613e258666d Type:ContainerStarted Data:d768276def4b37ab70feeee6455eb37f4fe6716bad6df283af0049c5724973e7} Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:53.376261 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" event=&{ID:7c63c9be-d1ad-44b6-8b33-09af3ee314af Type:ContainerStarted Data:f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3} Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:53.376956 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" event=&{ID:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 Type:ContainerStarted Data:1ffc39284205644ee2d20c66b8df819acc6bda29c81ecc932ae2237e02a208e7} Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:53.378188 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" event=&{ID:a40bf20d-67a4-45e7-b701-ca5a75376c1f Type:ContainerStarted Data:bf8032a2230aa8777b5ec76b9ff9222eb0d8d4700ec6b91316ee8cd870968d80} Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:53.382269 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" event=&{ID:88745ab7-efe7-42c0-a2e8-75591518333d Type:ContainerStarted Data:14626e8a3347256c5a502841426c6a0bfa7993e08a5404798831b7d49a3dddf7} Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.408319223Z" level=info msg="Created container 1806ad111afb95a724d10bd8c46e3bd826a078fefa581add2dbef820b5ef1cb4: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/thanos-query" id=92197d1a-4e43-4f25-b27e-0b511cb5b7ca name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.409006828Z" level=info msg="Starting container: 1806ad111afb95a724d10bd8c46e3bd826a078fefa581add2dbef820b5ef1cb4" id=f69d7743-c212-4496-b553-484547fb6e66 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.492713808Z" level=info msg="Started container" PID=5182 containerID=1806ad111afb95a724d10bd8c46e3bd826a078fefa581add2dbef820b5ef1cb4 description=openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/thanos-query id=f69d7743-c212-4496-b553-484547fb6e66 name=/runtime.v1.RuntimeService/StartContainer sandboxID=f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3 Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.533107418Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=dbb36c50-4215-40f5-ab0c-d16505c97ed0 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.533327066Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=dbb36c50-4215-40f5-ab0c-d16505c97ed0 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.545113938Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=28d1e164-2b3c-4d59-940c-5078adbc3349 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.545428779Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=28d1e164-2b3c-4d59-940c-5078adbc3349 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.546632897Z" level=info msg="Creating container: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/oauth-proxy" id=ef64cac8-d533-4bdc-a0a1-5acf8a7aaa14 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.546785893Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.735415557Z" level=info msg="Created container 1d7597386b38676c039e5b6aac3f7c678b4af05333bf527578cbc167c1f3ec5c: openshift-ingress/router-default-5cf6bbdff9-tnffs/router" id=0fe92ec6-94cf-4a75-9e1e-9c407ee741c0 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.737657379Z" level=info msg="Starting container: 1d7597386b38676c039e5b6aac3f7c678b4af05333bf527578cbc167c1f3ec5c" id=a5537886-729b-4b3d-a085-9eb2736b21b2 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.842361568Z" level=info msg="Started container" PID=5263 containerID=1d7597386b38676c039e5b6aac3f7c678b4af05333bf527578cbc167c1f3ec5c description=openshift-ingress/router-default-5cf6bbdff9-tnffs/router id=a5537886-729b-4b3d-a085-9eb2736b21b2 name=/runtime.v1.RuntimeService/StartContainer sandboxID=d768276def4b37ab70feeee6455eb37f4fe6716bad6df283af0049c5724973e7 Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.923837153Z" level=info msg="Created container ff8e0555d55e91975ea01b26878c5b45f1de8d8d28aa191358f2ab55fccbc19b: openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh/prometheus-adapter" id=79cd50ca-13ea-4c82-acce-19ba3f0f4acc name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.952301534Z" level=info msg="Created container a92439f77001fbb754b6894a21e8f5cc53e532d705ed818cd99da13a6465a6ef: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/kube-rbac-proxy-main" id=cb6dee0f-88b1-449a-996c-29fd0c1dc385 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:53.955466 2171 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-monitoring/alertmanager-main-0] Jan 05 09:16:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:53.955523 2171 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.965204238Z" level=info msg="Created container 0e9aa74411f0cbe12125f1767253aa311244845770097efc2342f5f2213a5f99: openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn/prometheus-operator-admission-webhook" id=20443b6b-06aa-47f6-b793-437c962a4a6d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:53.965417081Z" level=info msg="Starting container: ff8e0555d55e91975ea01b26878c5b45f1de8d8d28aa191358f2ab55fccbc19b" id=2ab2c4a9-d09f-4bd4-bcaa-e9dcd2f8138c name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.016920 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/05c9a6ea-140a-49a2-bde5-f220b05aa252-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.016968 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxsm5\" (UniqueName: \"kubernetes.io/projected/05c9a6ea-140a-49a2-bde5-f220b05aa252-kube-api-access-kxsm5\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.017006 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/05c9a6ea-140a-49a2-bde5-f220b05aa252-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.017041 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-web-config\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.017074 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/05c9a6ea-140a-49a2-bde5-f220b05aa252-config-out\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.017112 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.017148 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-config-volume\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.017182 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.017219 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.017255 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/05c9a6ea-140a-49a2-bde5-f220b05aa252-tls-assets\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.017288 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05c9a6ea-140a-49a2-bde5-f220b05aa252-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.017327 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-proxy\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-main-proxy\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:54.024967803Z" level=info msg="Starting container: a92439f77001fbb754b6894a21e8f5cc53e532d705ed818cd99da13a6465a6ef" id=9eae3eeb-1968-4fb2-9e68-a5469fef2990 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:54.053501929Z" level=info msg="Starting container: 0e9aa74411f0cbe12125f1767253aa311244845770097efc2342f5f2213a5f99" id=ccd2cfc9-9bdf-4ec9-af82-eeaf10d9acfa name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:54.098849142Z" level=info msg="Started container" PID=5269 containerID=a92439f77001fbb754b6894a21e8f5cc53e532d705ed818cd99da13a6465a6ef description=openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/kube-rbac-proxy-main id=9eae3eeb-1968-4fb2-9e68-a5469fef2990 name=/runtime.v1.RuntimeService/StartContainer sandboxID=1ffc39284205644ee2d20c66b8df819acc6bda29c81ecc932ae2237e02a208e7 Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.119358 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.119435 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.119475 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/05c9a6ea-140a-49a2-bde5-f220b05aa252-tls-assets\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.119509 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05c9a6ea-140a-49a2-bde5-f220b05aa252-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.119547 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-proxy\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-main-proxy\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.119594 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/05c9a6ea-140a-49a2-bde5-f220b05aa252-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.119632 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-kxsm5\" (UniqueName: \"kubernetes.io/projected/05c9a6ea-140a-49a2-bde5-f220b05aa252-kube-api-access-kxsm5\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.119662 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-web-config\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.119699 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/05c9a6ea-140a-49a2-bde5-f220b05aa252-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.119737 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/05c9a6ea-140a-49a2-bde5-f220b05aa252-config-out\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.119789 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.119824 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-config-volume\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.139911 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.147380 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-config-volume\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.148939 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05c9a6ea-140a-49a2-bde5-f220b05aa252-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.149902 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/05c9a6ea-140a-49a2-bde5-f220b05aa252-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.150224 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/05c9a6ea-140a-49a2-bde5-f220b05aa252-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.150829 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/05c9a6ea-140a-49a2-bde5-f220b05aa252-config-out\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:54.151449791Z" level=info msg="Started container" PID=5296 containerID=ff8e0555d55e91975ea01b26878c5b45f1de8d8d28aa191358f2ab55fccbc19b description=openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh/prometheus-adapter id=2ab2c4a9-d09f-4bd4-bcaa-e9dcd2f8138c name=/runtime.v1.RuntimeService/StartContainer sandboxID=bf8032a2230aa8777b5ec76b9ff9222eb0d8d4700ec6b91316ee8cd870968d80 Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.159441 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.161620 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/05c9a6ea-140a-49a2-bde5-f220b05aa252-tls-assets\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:54.183339161Z" level=info msg="Started container" PID=5266 containerID=0e9aa74411f0cbe12125f1767253aa311244845770097efc2342f5f2213a5f99 description=openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn/prometheus-operator-admission-webhook id=ccd2cfc9-9bdf-4ec9-af82-eeaf10d9acfa name=/runtime.v1.RuntimeService/StartContainer sandboxID=14626e8a3347256c5a502841426c6a0bfa7993e08a5404798831b7d49a3dddf7 Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.190874 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-proxy\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-main-proxy\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.192868 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-web-config\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.194895 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.230077 2171 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-monitoring/prometheus-k8s-0] Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.230125 2171 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.259925 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxsm5\" (UniqueName: \"kubernetes.io/projected/05c9a6ea-140a-49a2-bde5-f220b05aa252-kube-api-access-kxsm5\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.350728 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.350799 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.350876 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.350919 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.350958 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98gp9\" (UniqueName: \"kubernetes.io/projected/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-kube-api-access-98gp9\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.351046 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-web-config\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.351089 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.351127 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-kube-etcd-client-certs\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-kube-etcd-client-certs\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.351200 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-proxy\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-proxy\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.351243 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.351324 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-config\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.351373 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.351460 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.351508 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.351570 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-config-out\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.351627 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.351687 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.351728 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.351785 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.453472 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.453527 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.453556 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-config-out\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.453590 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.453622 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.453653 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.453687 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.453727 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.453774 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.453815 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.453848 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-98gp9\" (UniqueName: \"kubernetes.io/projected/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-kube-api-access-98gp9\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.453880 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.453915 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-web-config\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.453953 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.453987 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-kube-etcd-client-certs\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-kube-etcd-client-certs\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.454025 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-proxy\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-proxy\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.454066 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-config\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.454102 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.454137 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.454500 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.455257 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.456550 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.516953 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.518561 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.527941 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-config-out\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.579053 2171 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.580248 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.580648 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.581372 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.584071 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:54.592836562Z" level=info msg="Running pod sandbox: openshift-monitoring/alertmanager-main-0/POD" id=c37bad3e-1175-4fa6-a344-e8a02c3c7938 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:54.592906662Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.606653 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-kube-etcd-client-certs\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-kube-etcd-client-certs\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.607098 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.607490 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-web-config\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.621237 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.664221 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-config\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.664818 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-98gp9\" (UniqueName: \"kubernetes.io/projected/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-kube-api-access-98gp9\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.675351 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-proxy\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-proxy\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.678064 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.678476 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:54.669496383Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=05c26d14-54d3-4fba-b66e-8d223e8cf6e6 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:54.669719614Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=05c26d14-54d3-4fba-b66e-8d223e8cf6e6 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.714041 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" event=&{ID:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 Type:ContainerStarted Data:a92439f77001fbb754b6894a21e8f5cc53e532d705ed818cd99da13a6465a6ef} Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.715375 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" event=&{ID:a40bf20d-67a4-45e7-b701-ca5a75376c1f Type:ContainerStarted Data:ff8e0555d55e91975ea01b26878c5b45f1de8d8d28aa191358f2ab55fccbc19b} Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.716675 2171 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.717879 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" event=&{ID:88745ab7-efe7-42c0-a2e8-75591518333d Type:ContainerStarted Data:0e9aa74411f0cbe12125f1767253aa311244845770097efc2342f5f2213a5f99} Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.718878 2171 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.719991 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" event=&{ID:c3b10ff4-a243-4a4e-b27d-f613e258666d Type:ContainerStarted Data:1d7597386b38676c039e5b6aac3f7c678b4af05333bf527578cbc167c1f3ec5c} Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.722075 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" event=&{ID:7c63c9be-d1ad-44b6-8b33-09af3ee314af Type:ContainerStarted Data:1806ad111afb95a724d10bd8c46e3bd826a078fefa581add2dbef820b5ef1cb4} Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.728197 2171 patch_prober.go:29] interesting pod/prometheus-operator-admission-webhook-598468bb49-94dsn container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.129.2.10:8443/healthz\": dial tcp 10.129.2.10:8443: connect: connection refused" start-of-body= Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.728260 2171 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" podUID=88745ab7-efe7-42c0-a2e8-75591518333d containerName="prometheus-operator-admission-webhook" probeResult=failure output="Get \"https://10.129.2.10:8443/healthz\": dial tcp 10.129.2.10:8443: connect: connection refused" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:54.761007 2171 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:54.778033128Z" level=info msg="Running pod sandbox: openshift-monitoring/prometheus-k8s-0/POD" id=7930e7ba-9058-496c-ac9d-6e4b17af6c57 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:54.778097737Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:54.778626308Z" level=info msg="Created container 5b60f1d912768234f658c95fc37c5638dd28e2bbc01c64bb2abb830967916a74: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/oauth-proxy" id=ef64cac8-d533-4bdc-a0a1-5acf8a7aaa14 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:54.834601651Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=5f383a24-445e-4ffe-8d2f-0eb0058471e4 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:54.834851835Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=5f383a24-445e-4ffe-8d2f-0eb0058471e4 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:54.835094718Z" level=info msg="Starting container: 5b60f1d912768234f658c95fc37c5638dd28e2bbc01c64bb2abb830967916a74" id=e8eb10fa-5505-4afb-a160-e4ff86f495aa name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:54.917741387Z" level=info msg="Creating container: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/kube-rbac-proxy-self" id=1ccf41a2-408a-4304-bb1b-180eb297e1c2 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:54.917928455Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:55.019859281Z" level=info msg="Started container" PID=5359 containerID=5b60f1d912768234f658c95fc37c5638dd28e2bbc01c64bb2abb830967916a74 description=openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/oauth-proxy id=e8eb10fa-5505-4afb-a160-e4ff86f495aa name=/runtime.v1.RuntimeService/StartContainer sandboxID=f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3 Jan 05 09:16:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:55.254599299Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=49ddd26a-88a6-416b-88d3-26a03b3b79e1 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:55.254835545Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=49ddd26a-88a6-416b-88d3-26a03b3b79e1 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:55.276107499Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=8d240a72-a527-49fa-881f-d79217aca335 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:55.276329768Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=8d240a72-a527-49fa-881f-d79217aca335 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:55.318131750Z" level=info msg="Creating container: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy" id=c1791e2a-bc5b-43f5-94db-c79667bfd7aa name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:55.318269284Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:55.699924 2171 kubelet.go:2229] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:16:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:55.708969 2171 kubelet.go:2229] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:16:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:55.780644 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" event=&{ID:7c63c9be-d1ad-44b6-8b33-09af3ee314af Type:ContainerStarted Data:5b60f1d912768234f658c95fc37c5638dd28e2bbc01c64bb2abb830967916a74} Jan 05 09:16:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:55.780859 2171 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:16:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:55.795697 2171 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:16:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:55.797172 2171 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" Jan 05 09:16:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:56.586982530Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:85f671344a3bc1584db4caed39939b6dd55644a5bd806e118527f31dac91252e" id=0c449ce7-8276-4cfa-8fef-4be300f09bba name=/runtime.v1.ImageService/PullImage Jan 05 09:16:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:56.589107290Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e5521a3d1f64483e63ffe5bc23d9f882781cd4d152ac6de72013eaefe6dfcf89" id=0f68ee03-c88b-45c9-8459-2e255fce5cd2 name=/runtime.v1.ImageService/PullImage Jan 05 09:16:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:56.599147458Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e5521a3d1f64483e63ffe5bc23d9f882781cd4d152ac6de72013eaefe6dfcf89" id=f78f4573-4fcd-406f-a5d8-9078cc501619 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:56.599661560Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:85f671344a3bc1584db4caed39939b6dd55644a5bd806e118527f31dac91252e" id=9ff7ff6c-7bc0-4701-8274-939443427c8d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:56.604488643Z" level=info msg="Got pod network &{Name:alertmanager-main-0 Namespace:openshift-monitoring ID:0bb16724854db8df010963df02e5853c885cede9ad0447f86da449f730ebabf0 UID:05c9a6ea-140a-49a2-bde5-f220b05aa252 NetNS:/var/run/netns/a88b2c9c-80dc-4225-a81d-ec8a37f3cf76 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:16:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:56.604520865Z" level=info msg="Adding pod openshift-monitoring_alertmanager-main-0 to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:16:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:56.654662642Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:af30ce06a0837328332ff57ed27c48095a8ed9600535c919aa69fa76524919c6,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e5521a3d1f64483e63ffe5bc23d9f882781cd4d152ac6de72013eaefe6dfcf89],Size_:371510113,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=f78f4573-4fcd-406f-a5d8-9078cc501619 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:56.656479379Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b66b2c4519cdd77137d5eaa713199fe202373c0d1e029b31472b9fc8a68a2aa7,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:85f671344a3bc1584db4caed39939b6dd55644a5bd806e118527f31dac91252e],Size_:338794817,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=9ff7ff6c-7bc0-4701-8274-939443427c8d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:56.659692945Z" level=info msg="Got pod network &{Name:prometheus-k8s-0 Namespace:openshift-monitoring ID:33e5ecc33cac76d9879725e41d559c60e7711f7e5422fe607edbdadb5460eb11 UID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 NetNS:/var/run/netns/6bbe39cf-79d5-4f40-b2bc-384feaa91c56 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:16:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:56.659720288Z" level=info msg="Adding pod openshift-monitoring_prometheus-k8s-0 to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:16:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:56.667482837Z" level=info msg="Creating container: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-state-metrics" id=b63ff050-4c08-4438-837b-b0ab934da78a name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:56.667665475Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:56.672461833Z" level=info msg="Creating container: openshift-monitoring/telemeter-client-5598c645c4-mftdv/telemeter-client" id=2ad82c2c-d96e-4221-a95f-b5728230303f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:56.672561359Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:56.979483487Z" level=info msg="Created container 972cda72e60cf7a211036a270be6909a83c8720dc668efa2e2dd303e21789191: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/kube-rbac-proxy-self" id=1ccf41a2-408a-4304-bb1b-180eb297e1c2 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:56.985467493Z" level=info msg="Starting container: 972cda72e60cf7a211036a270be6909a83c8720dc668efa2e2dd303e21789191" id=e9263260-136a-4c92-a9b0-0f17b3f7fee9 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.021216305Z" level=info msg="Started container" PID=5710 containerID=972cda72e60cf7a211036a270be6909a83c8720dc668efa2e2dd303e21789191 description=openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/kube-rbac-proxy-self id=e9263260-136a-4c92-a9b0-0f17b3f7fee9 name=/runtime.v1.RuntimeService/StartContainer sandboxID=1ffc39284205644ee2d20c66b8df819acc6bda29c81ecc932ae2237e02a208e7 Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.063734461Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:78252776c28ff0ef5195173aebd633dc3a71e5d4732ca01f9104bb0aab07570e" id=2a672f10-c1fc-449f-a67e-2fa8990b63cb name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.063985973Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:78252776c28ff0ef5195173aebd633dc3a71e5d4732ca01f9104bb0aab07570e not found" id=2a672f10-c1fc-449f-a67e-2fa8990b63cb name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.067184336Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:78252776c28ff0ef5195173aebd633dc3a71e5d4732ca01f9104bb0aab07570e" id=e7df918b-006e-4361-95d4-2cd54a38703f name=/runtime.v1.ImageService/PullImage Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.068417774Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:78252776c28ff0ef5195173aebd633dc3a71e5d4732ca01f9104bb0aab07570e\"" Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.132226115Z" level=info msg="Created container 0ba3e4783da8b547347f29221e83019680bad22ec9f3f891d05c7beed9fcbcad: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy" id=c1791e2a-bc5b-43f5-94db-c79667bfd7aa name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.132870029Z" level=info msg="Starting container: 0ba3e4783da8b547347f29221e83019680bad22ec9f3f891d05c7beed9fcbcad" id=1284527c-58d5-4b4b-8732-3420e1cd07de name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.161596962Z" level=info msg="Started container" PID=5752 containerID=0ba3e4783da8b547347f29221e83019680bad22ec9f3f891d05c7beed9fcbcad description=openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy id=1284527c-58d5-4b4b-8732-3420e1cd07de name=/runtime.v1.RuntimeService/StartContainer sandboxID=f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3 Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.202850134Z" level=info msg="Created container a64a8eec7c21e1dc010077a81f71bfd1ef40287e87f8d9147af65ba664e54322: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-state-metrics" id=b63ff050-4c08-4438-837b-b0ab934da78a name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.204123209Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816" id=77e02e9a-9dbd-4d85-8ff4-13dbad5ca657 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.204353807Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:955ee2b41ecd88e3a9c0183a0b65d6edbf524cefa1318837dcbc82b0466c4d55,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816],Size_:309576938,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=77e02e9a-9dbd-4d85-8ff4-13dbad5ca657 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.204480737Z" level=info msg="Starting container: a64a8eec7c21e1dc010077a81f71bfd1ef40287e87f8d9147af65ba664e54322" id=7299146a-872b-4078-add6-25cff01681db name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.239110205Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816" id=790de141-e232-4a2a-af11-21cbdac2cbef name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.239340471Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:955ee2b41ecd88e3a9c0183a0b65d6edbf524cefa1318837dcbc82b0466c4d55,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816],Size_:309576938,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=790de141-e232-4a2a-af11-21cbdac2cbef name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.240552322Z" level=info msg="Creating container: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/prom-label-proxy" id=39d838eb-73bd-4a91-bf89-4d4ee1c70e51 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.240668925Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.243201326Z" level=info msg="Started container" PID=5746 containerID=a64a8eec7c21e1dc010077a81f71bfd1ef40287e87f8d9147af65ba664e54322 description=openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-state-metrics id=7299146a-872b-4078-add6-25cff01681db name=/runtime.v1.RuntimeService/StartContainer sandboxID=bb417fcfe195906c7b26f0d9c20687f555b79c545c37fe3e4cebd9d60e23d18c Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.266193369Z" level=info msg="Created container 183543c83fed2e18744fbae52d6331fb21aa60a9b2756ab8cb9cfbc9fdfbad3f: openshift-monitoring/telemeter-client-5598c645c4-mftdv/telemeter-client" id=2ad82c2c-d96e-4221-a95f-b5728230303f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.266662458Z" level=info msg="Starting container: 183543c83fed2e18744fbae52d6331fb21aa60a9b2756ab8cb9cfbc9fdfbad3f" id=2eaa61aa-af57-40d7-becb-6e9d79b6a63e name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.301598418Z" level=info msg="Started container" PID=5776 containerID=183543c83fed2e18744fbae52d6331fb21aa60a9b2756ab8cb9cfbc9fdfbad3f description=openshift-monitoring/telemeter-client-5598c645c4-mftdv/telemeter-client id=2eaa61aa-af57-40d7-becb-6e9d79b6a63e name=/runtime.v1.RuntimeService/StartContainer sandboxID=0a5caa2be3488b9c8c36d43ddc5d5a1784a28176b83143a0b5296da42812377f Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.302439935Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=e93c1fe7-90de-4b92-a9c6-fb58aeabf763 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.323321207Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=e93c1fe7-90de-4b92-a9c6-fb58aeabf763 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.330419176Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:78252776c28ff0ef5195173aebd633dc3a71e5d4732ca01f9104bb0aab07570e\"" Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.349444570Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=d7487a7c-88d0-4284-8696-994452620ae9 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.349645111Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=d7487a7c-88d0-4284-8696-994452620ae9 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.351871679Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=655ef8fe-160c-4c7a-a0ba-6e02d5a8590b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.352048479Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=655ef8fe-160c-4c7a-a0ba-6e02d5a8590b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.352120854Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=52512022-b543-4cff-8455-667398ea7159 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.352287088Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=52512022-b543-4cff-8455-667398ea7159 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.359846741Z" level=info msg="Creating container: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-rbac-proxy-main" id=240ba78c-6645-44ad-936e-9b9b487529b4 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.359949842Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.363207667Z" level=info msg="Creating container: openshift-monitoring/telemeter-client-5598c645c4-mftdv/reload" id=337cc115-ccaa-4d2d-80ed-a79eba5d1b91 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.363304007Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.710858974Z" level=info msg="Created container 0e5bb1b82e50570e1238f090a0c69c8ce8c4924181c9fe1d3c2e79bbcaf04cf7: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/prom-label-proxy" id=39d838eb-73bd-4a91-bf89-4d4ee1c70e51 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.711272410Z" level=info msg="Starting container: 0e5bb1b82e50570e1238f090a0c69c8ce8c4924181c9fe1d3c2e79bbcaf04cf7" id=64f06b22-937d-4116-be7e-74906b9a0d9c name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.741718764Z" level=info msg="Started container" PID=5894 containerID=0e5bb1b82e50570e1238f090a0c69c8ce8c4924181c9fe1d3c2e79bbcaf04cf7 description=openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/prom-label-proxy id=64f06b22-937d-4116-be7e-74906b9a0d9c name=/runtime.v1.RuntimeService/StartContainer sandboxID=f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3 Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.781351742Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=af6ff2ee-64c0-4273-912c-6ee7d8cd0063 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.781592469Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=af6ff2ee-64c0-4273-912c-6ee7d8cd0063 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.782564530Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=5e0b02bf-4202-4f3a-bb9e-368882edc46a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.782798092Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=5e0b02bf-4202-4f3a-bb9e-368882edc46a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.783898899Z" level=info msg="Creating container: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy-rules" id=55df2da3-eafd-41b7-a5c2-81c1b72fc6f2 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:57.784012025Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:57.928481 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" event=&{ID:7c63c9be-d1ad-44b6-8b33-09af3ee314af Type:ContainerStarted Data:0e5bb1b82e50570e1238f090a0c69c8ce8c4924181c9fe1d3c2e79bbcaf04cf7} Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:57.928520 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" event=&{ID:7c63c9be-d1ad-44b6-8b33-09af3ee314af Type:ContainerStarted Data:0ba3e4783da8b547347f29221e83019680bad22ec9f3f891d05c7beed9fcbcad} Jan 05 09:16:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:57.953521 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" event=&{ID:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 Type:ContainerStarted Data:972cda72e60cf7a211036a270be6909a83c8720dc668efa2e2dd303e21789191} Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:58.013765 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" event=&{ID:fc45e847-8f55-4c48-8c48-4bfee2419dca Type:ContainerStarted Data:183543c83fed2e18744fbae52d6331fb21aa60a9b2756ab8cb9cfbc9fdfbad3f} Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:58.016645 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" event=&{ID:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e Type:ContainerStarted Data:a64a8eec7c21e1dc010077a81f71bfd1ef40287e87f8d9147af65ba664e54322} Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.087637899Z" level=info msg="Created container 509279176d1ba7a99db50371d2e6fbd2aa072937e3210855121bfc88ffaf8250: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-rbac-proxy-main" id=240ba78c-6645-44ad-936e-9b9b487529b4 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.127912714Z" level=info msg="Starting container: 509279176d1ba7a99db50371d2e6fbd2aa072937e3210855121bfc88ffaf8250" id=664f4c25-cbf5-4e75-a2ac-4e6505788b4d name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.186179955Z" level=info msg="Created container 14af629856736bb66f0a281297394a8dac31178dfcdb2788d39f3744b283ba3e: openshift-monitoring/telemeter-client-5598c645c4-mftdv/reload" id=337cc115-ccaa-4d2d-80ed-a79eba5d1b91 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.203872565Z" level=info msg="Starting container: 14af629856736bb66f0a281297394a8dac31178dfcdb2788d39f3744b283ba3e" id=19df5e7c-037e-4ec2-9987-ccc72295e0ad name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.210038850Z" level=info msg="Started container" PID=5941 containerID=509279176d1ba7a99db50371d2e6fbd2aa072937e3210855121bfc88ffaf8250 description=openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-rbac-proxy-main id=664f4c25-cbf5-4e75-a2ac-4e6505788b4d name=/runtime.v1.RuntimeService/StartContainer sandboxID=bb417fcfe195906c7b26f0d9c20687f555b79c545c37fe3e4cebd9d60e23d18c Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.265039031Z" level=info msg="Started container" PID=5949 containerID=14af629856736bb66f0a281297394a8dac31178dfcdb2788d39f3744b283ba3e description=openshift-monitoring/telemeter-client-5598c645c4-mftdv/reload id=19df5e7c-037e-4ec2-9987-ccc72295e0ad name=/runtime.v1.RuntimeService/StartContainer sandboxID=0a5caa2be3488b9c8c36d43ddc5d5a1784a28176b83143a0b5296da42812377f Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:58.268005 2171 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/prometheus-k8s-0] Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: 2023-01-05T09:16:58Z [verbose] Add: openshift-monitoring:prometheus-k8s-0:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/6bbe39cf-79d5-4f40-b2bc-384feaa91c56"}],"ips":[{"version":"4","interface":0,"address":"10.129.2.14/23","gateway":"10.129.2.1"}],"dns":{}} Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: I0105 09:16:58.238214 5683 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"prometheus-k8s-0", UID:"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879", APIVersion:"v1", ResourceVersion:"80030", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.129.2.14/23] from openshift-sdn Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.278859548Z" level=info msg="Got pod network &{Name:prometheus-k8s-0 Namespace:openshift-monitoring ID:33e5ecc33cac76d9879725e41d559c60e7711f7e5422fe607edbdadb5460eb11 UID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 NetNS:/var/run/netns/6bbe39cf-79d5-4f40-b2bc-384feaa91c56 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.279003657Z" level=info msg="Checking pod openshift-monitoring_prometheus-k8s-0 for CNI network multus-cni-network (type=multus)" Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.304920573Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=c69c55cc-641f-4b9d-9dea-25b32cc3af8d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.305145547Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=c69c55cc-641f-4b9d-9dea-25b32cc3af8d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.311267034Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=baeefef1-640b-4cad-bfe1-25ff979227cb name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.311626327Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=baeefef1-640b-4cad-bfe1-25ff979227cb name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.315549044Z" level=info msg="Creating container: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-rbac-proxy-self" id=4ffbbfb7-c7fc-4801-8bb5-82bff19d400f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.315665867Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.328989920Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=91a7c443-9633-426d-beb7-838e5a98e389 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.330360987Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=91a7c443-9633-426d-beb7-838e5a98e389 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.331331291Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=0397baef-362b-4120-9a9b-d2030876bb46 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.331539253Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=0397baef-362b-4120-9a9b-d2030876bb46 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.333805522Z" level=info msg="Creating container: openshift-monitoring/telemeter-client-5598c645c4-mftdv/kube-rbac-proxy" id=fc2676d6-33fd-4d99-9ed6-41005373a80c name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.333916044Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.371025700Z" level=info msg="Ran pod sandbox 33e5ecc33cac76d9879725e41d559c60e7711f7e5422fe607edbdadb5460eb11 with infra container: openshift-monitoring/prometheus-k8s-0/POD" id=7930e7ba-9058-496c-ac9d-6e4b17af6c57 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.398370901Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=49243663-91d9-4580-8293-5c1d1a621fba name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.398584528Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=49243663-91d9-4580-8293-5c1d1a621fba name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.406428437Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=63f85ea8-4d17-43d5-9b23-7e4019380506 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.406630130Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=63f85ea8-4d17-43d5-9b23-7e4019380506 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.414710187Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-0/init-config-reloader" id=1330c267-c58b-4220-8a08-99d5bfcc10d7 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.414998359Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.686156234Z" level=info msg="Created container d14df105d1078625422b2a59ff6a8bdadf9cd80f1516885c6074d5a7d6d78616: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy-rules" id=55df2da3-eafd-41b7-a5c2-81c1b72fc6f2 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.687065743Z" level=info msg="Starting container: d14df105d1078625422b2a59ff6a8bdadf9cd80f1516885c6074d5a7d6d78616" id=b4b037b4-e152-440f-bd14-5ec61751d82d name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.699952218Z" level=info msg="Started container" PID=6037 containerID=d14df105d1078625422b2a59ff6a8bdadf9cd80f1516885c6074d5a7d6d78616 description=openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy-rules id=b4b037b4-e152-440f-bd14-5ec61751d82d name=/runtime.v1.RuntimeService/StartContainer sandboxID=f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3 Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.748162805Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=382501ab-9398-44ab-9fdf-858e9c1e8430 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.748331559Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=382501ab-9398-44ab-9fdf-858e9c1e8430 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.749363704Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=b49c0379-44c5-4d07-b740-bc2dc43838fe name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.749606624Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b49c0379-44c5-4d07-b740-bc2dc43838fe name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.750648603Z" level=info msg="Creating container: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy-metrics" id=916d1ef1-1ce6-43ad-9d39-10597b8a2017 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.750791824Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.941813695Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:78252776c28ff0ef5195173aebd633dc3a71e5d4732ca01f9104bb0aab07570e" id=e7df918b-006e-4361-95d4-2cd54a38703f name=/runtime.v1.ImageService/PullImage Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.951121805Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:78252776c28ff0ef5195173aebd633dc3a71e5d4732ca01f9104bb0aab07570e" id=ce061bfc-157d-4b53-be63-2665a02567a0 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.953355947Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:ab19b64ab85c6b98c0381b09c3bd6f6db37bb31b25f12aa07bf856e8be07eac7,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:78252776c28ff0ef5195173aebd633dc3a71e5d4732ca01f9104bb0aab07570e],Size_:365012751,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=ce061bfc-157d-4b53-be63-2665a02567a0 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.954472862Z" level=info msg="Creating container: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/openshift-state-metrics" id=fc33bede-8ff4-4fa8-bd16-84755f505789 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.954577947Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:58.995278 2171 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/alertmanager-main-0] Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: 2023-01-05T09:16:58Z [verbose] Add: openshift-monitoring:alertmanager-main-0:05c9a6ea-140a-49a2-bde5-f220b05aa252:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/a88b2c9c-80dc-4225-a81d-ec8a37f3cf76"}],"ips":[{"version":"4","interface":0,"address":"10.129.2.13/23","gateway":"10.129.2.1"}],"dns":{}} Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: I0105 09:16:58.963905 5673 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"alertmanager-main-0", UID:"05c9a6ea-140a-49a2-bde5-f220b05aa252", APIVersion:"v1", ResourceVersion:"80012", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.129.2.13/23] from openshift-sdn Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.997641489Z" level=info msg="Got pod network &{Name:alertmanager-main-0 Namespace:openshift-monitoring ID:0bb16724854db8df010963df02e5853c885cede9ad0447f86da449f730ebabf0 UID:05c9a6ea-140a-49a2-bde5-f220b05aa252 NetNS:/var/run/netns/a88b2c9c-80dc-4225-a81d-ec8a37f3cf76 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:16:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:58.997808504Z" level=info msg="Checking pod openshift-monitoring_alertmanager-main-0 for CNI network multus-cni-network (type=multus)" Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 09:16:59.001146 2171 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05c9a6ea_140a_49a2_bde5_f220b05aa252.slice/crio-0bb16724854db8df010963df02e5853c885cede9ad0447f86da449f730ebabf0.scope WatchSource:0}: Error finding container 0bb16724854db8df010963df02e5853c885cede9ad0447f86da449f730ebabf0: Status 404 returned error can't find the container with id 0bb16724854db8df010963df02e5853c885cede9ad0447f86da449f730ebabf0 Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.020359882Z" level=info msg="Ran pod sandbox 0bb16724854db8df010963df02e5853c885cede9ad0447f86da449f730ebabf0 with infra container: openshift-monitoring/alertmanager-main-0/POD" id=c37bad3e-1175-4fa6-a344-e8a02c3c7938 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.043635767Z" level=info msg="Created container 5c8fab5cf2da97490127bff84313a78ffc021feac020252239d18a75f21cbc6b: openshift-monitoring/telemeter-client-5598c645c4-mftdv/kube-rbac-proxy" id=fc2676d6-33fd-4d99-9ed6-41005373a80c name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.044987531Z" level=info msg="Starting container: 5c8fab5cf2da97490127bff84313a78ffc021feac020252239d18a75f21cbc6b" id=b95aad35-9558-4081-93b0-1777832934a3 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.045187430Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e6173b5d77def9e8058c7ff82f1ad65f537d1a81f88fcdabe6d8e30bff49ca89" id=6ef2fbac-663c-4850-be4c-273ba45dedc0 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.045456018Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:d168c738461c25cc0c3fa96cfcdf2dc337ca08f5bf3a6b550fcb29a59ea39270,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e6173b5d77def9e8058c7ff82f1ad65f537d1a81f88fcdabe6d8e30bff49ca89],Size_:333780768,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=6ef2fbac-663c-4850-be4c-273ba45dedc0 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:59.057136 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event=&{ID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Type:ContainerStarted Data:33e5ecc33cac76d9879725e41d559c60e7711f7e5422fe607edbdadb5460eb11} Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.057378525Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e6173b5d77def9e8058c7ff82f1ad65f537d1a81f88fcdabe6d8e30bff49ca89" id=ca576eee-7012-42d0-aadc-40cf027b3f3f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.057613656Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:d168c738461c25cc0c3fa96cfcdf2dc337ca08f5bf3a6b550fcb29a59ea39270,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e6173b5d77def9e8058c7ff82f1ad65f537d1a81f88fcdabe6d8e30bff49ca89],Size_:333780768,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=ca576eee-7012-42d0-aadc-40cf027b3f3f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.069368720Z" level=info msg="Creating container: openshift-monitoring/alertmanager-main-0/alertmanager" id=3cefb724-e2f7-476a-8f7a-698da36ba790 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.069533380Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:59.078506 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" event=&{ID:fc45e847-8f55-4c48-8c48-4bfee2419dca Type:ContainerStarted Data:14af629856736bb66f0a281297394a8dac31178dfcdb2788d39f3744b283ba3e} Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:59.088237 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" event=&{ID:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e Type:ContainerStarted Data:509279176d1ba7a99db50371d2e6fbd2aa072937e3210855121bfc88ffaf8250} Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:59.089238 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event=&{ID:05c9a6ea-140a-49a2-bde5-f220b05aa252 Type:ContainerStarted Data:0bb16724854db8df010963df02e5853c885cede9ad0447f86da449f730ebabf0} Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:16:59.091539 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" event=&{ID:7c63c9be-d1ad-44b6-8b33-09af3ee314af Type:ContainerStarted Data:d14df105d1078625422b2a59ff6a8bdadf9cd80f1516885c6074d5a7d6d78616} Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.128541638Z" level=info msg="Started container" PID=6084 containerID=5c8fab5cf2da97490127bff84313a78ffc021feac020252239d18a75f21cbc6b description=openshift-monitoring/telemeter-client-5598c645c4-mftdv/kube-rbac-proxy id=b95aad35-9558-4081-93b0-1777832934a3 name=/runtime.v1.RuntimeService/StartContainer sandboxID=0a5caa2be3488b9c8c36d43ddc5d5a1784a28176b83143a0b5296da42812377f Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.145777052Z" level=info msg="Created container 2dca518d6d28aeafb7a13c3ccc01e3f5af14216c4354b60ce83f720c1064b4b7: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-rbac-proxy-self" id=4ffbbfb7-c7fc-4801-8bb5-82bff19d400f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.146338031Z" level=info msg="Starting container: 2dca518d6d28aeafb7a13c3ccc01e3f5af14216c4354b60ce83f720c1064b4b7" id=9bf507ab-050e-4433-a611-4beaa91319c4 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.196718355Z" level=info msg="Started container" PID=6091 containerID=2dca518d6d28aeafb7a13c3ccc01e3f5af14216c4354b60ce83f720c1064b4b7 description=openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-rbac-proxy-self id=9bf507ab-050e-4433-a611-4beaa91319c4 name=/runtime.v1.RuntimeService/StartContainer sandboxID=bb417fcfe195906c7b26f0d9c20687f555b79c545c37fe3e4cebd9d60e23d18c Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.282649636Z" level=info msg="Created container 87856f0f10585dd547492ed5711f26f3409ec28f07c808cd58bb845c8f6d7522: openshift-monitoring/prometheus-k8s-0/init-config-reloader" id=1330c267-c58b-4220-8a08-99d5bfcc10d7 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.285477546Z" level=info msg="Starting container: 87856f0f10585dd547492ed5711f26f3409ec28f07c808cd58bb845c8f6d7522" id=7a993ec0-3362-4765-b78c-f513ff76bb21 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.301047972Z" level=info msg="Created container 722bd8448df2ddebd63be975ab1961567271a1f1cca344e95f989bcc07d89d6e: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy-metrics" id=916d1ef1-1ce6-43ad-9d39-10597b8a2017 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.301200995Z" level=info msg="Created container 30b2d68c19c82a725bb94bba811fe395846878cd0d20034d15397fe68fbe2865: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/openshift-state-metrics" id=fc33bede-8ff4-4fa8-bd16-84755f505789 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.301705968Z" level=info msg="Starting container: 722bd8448df2ddebd63be975ab1961567271a1f1cca344e95f989bcc07d89d6e" id=35220bbc-2a11-4287-a17b-3dbc3e3a1b15 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.301937278Z" level=info msg="Starting container: 30b2d68c19c82a725bb94bba811fe395846878cd0d20034d15397fe68fbe2865" id=eeb81e95-8631-467e-942e-d489678e8b9d name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.324062124Z" level=info msg="Started container" PID=6152 containerID=722bd8448df2ddebd63be975ab1961567271a1f1cca344e95f989bcc07d89d6e description=openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy-metrics id=35220bbc-2a11-4287-a17b-3dbc3e3a1b15 name=/runtime.v1.RuntimeService/StartContainer sandboxID=f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3 Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.336171850Z" level=info msg="Started container" PID=6163 containerID=87856f0f10585dd547492ed5711f26f3409ec28f07c808cd58bb845c8f6d7522 description=openshift-monitoring/prometheus-k8s-0/init-config-reloader id=7a993ec0-3362-4765-b78c-f513ff76bb21 name=/runtime.v1.RuntimeService/StartContainer sandboxID=33e5ecc33cac76d9879725e41d559c60e7711f7e5422fe607edbdadb5460eb11 Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.354024867Z" level=info msg="Started container" PID=6153 containerID=30b2d68c19c82a725bb94bba811fe395846878cd0d20034d15397fe68fbe2865 description=openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/openshift-state-metrics id=eeb81e95-8631-467e-942e-d489678e8b9d name=/runtime.v1.RuntimeService/StartContainer sandboxID=1ffc39284205644ee2d20c66b8df819acc6bda29c81ecc932ae2237e02a208e7 Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.541432705Z" level=info msg="Created container 006114a568589213245ba8263ebfa77e665fa4cac909846666c0dede6b7512f9: openshift-monitoring/alertmanager-main-0/alertmanager" id=3cefb724-e2f7-476a-8f7a-698da36ba790 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.542059520Z" level=info msg="Starting container: 006114a568589213245ba8263ebfa77e665fa4cac909846666c0dede6b7512f9" id=5b4858d2-ac45-47d3-9d70-34199e3d9532 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.556904815Z" level=info msg="Started container" PID=6273 containerID=006114a568589213245ba8263ebfa77e665fa4cac909846666c0dede6b7512f9 description=openshift-monitoring/alertmanager-main-0/alertmanager id=5b4858d2-ac45-47d3-9d70-34199e3d9532 name=/runtime.v1.RuntimeService/StartContainer sandboxID=0bb16724854db8df010963df02e5853c885cede9ad0447f86da449f730ebabf0 Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.572576837Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=8ad01cde-2f9a-4107-9ddf-2df2736aa9af name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.572806623Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=8ad01cde-2f9a-4107-9ddf-2df2736aa9af name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.576095426Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=041f0384-4b28-40c5-879b-7f0c810024d2 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.576278769Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=041f0384-4b28-40c5-879b-7f0c810024d2 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.577142516Z" level=info msg="Creating container: openshift-monitoring/alertmanager-main-0/config-reloader" id=e11936b8-b3cd-4a30-8409-ea56ea7f6976 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.577247185Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.712328865Z" level=info msg="Created container 744edbc686210b5388467d89b22a8f7cad5ce589c1225d191c8e1a4a5f5b4413: openshift-monitoring/alertmanager-main-0/config-reloader" id=e11936b8-b3cd-4a30-8409-ea56ea7f6976 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.712821373Z" level=info msg="Starting container: 744edbc686210b5388467d89b22a8f7cad5ce589c1225d191c8e1a4a5f5b4413" id=2ff36313-0621-4d7e-a46c-2457db9fe5ff name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.720570014Z" level=info msg="Started container" PID=6338 containerID=744edbc686210b5388467d89b22a8f7cad5ce589c1225d191c8e1a4a5f5b4413 description=openshift-monitoring/alertmanager-main-0/config-reloader id=2ff36313-0621-4d7e-a46c-2457db9fe5ff name=/runtime.v1.RuntimeService/StartContainer sandboxID=0bb16724854db8df010963df02e5853c885cede9ad0447f86da449f730ebabf0 Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.731283546Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=a4f4ac5c-04e8-4b7e-992e-0edd52ffff24 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.731491567Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=a4f4ac5c-04e8-4b7e-992e-0edd52ffff24 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.732237546Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=ee841a48-2ec5-412b-822f-2080c69d2f8e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.732433395Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=ee841a48-2ec5-412b-822f-2080c69d2f8e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.733293426Z" level=info msg="Creating container: openshift-monitoring/alertmanager-main-0/alertmanager-proxy" id=7e3da0b8-e841-4d6b-a0d7-21e4c0fce06d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.733425678Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.845566926Z" level=info msg="Created container 43228eea868ea1a6b603b81a7625cda951823b52f39f51806c1ba0ccdccbda56: openshift-monitoring/alertmanager-main-0/alertmanager-proxy" id=7e3da0b8-e841-4d6b-a0d7-21e4c0fce06d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.846206182Z" level=info msg="Starting container: 43228eea868ea1a6b603b81a7625cda951823b52f39f51806c1ba0ccdccbda56" id=9d386cb3-8078-40cd-8998-ec945057b5ca name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.853704885Z" level=info msg="Started container" PID=6374 containerID=43228eea868ea1a6b603b81a7625cda951823b52f39f51806c1ba0ccdccbda56 description=openshift-monitoring/alertmanager-main-0/alertmanager-proxy id=9d386cb3-8078-40cd-8998-ec945057b5ca name=/runtime.v1.RuntimeService/StartContainer sandboxID=0bb16724854db8df010963df02e5853c885cede9ad0447f86da449f730ebabf0 Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.865717677Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=f4c2b6dd-5394-46c4-93fc-718967ea9694 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.865957501Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=f4c2b6dd-5394-46c4-93fc-718967ea9694 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.866999936Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=e6f8ea1f-f508-4dec-8ff3-42fbb01a4921 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.867186429Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=e6f8ea1f-f508-4dec-8ff3-42fbb01a4921 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.868349341Z" level=info msg="Creating container: openshift-monitoring/alertmanager-main-0/kube-rbac-proxy" id=14eecf6a-2d2c-4dfa-9638-99ad37703535 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.868501750Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.986910208Z" level=info msg="Created container 7750d5fb0eed5aadcfb300d954cd8a06c38d92001a4babb24681e14257ee1005: openshift-monitoring/alertmanager-main-0/kube-rbac-proxy" id=14eecf6a-2d2c-4dfa-9638-99ad37703535 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.987316358Z" level=info msg="Starting container: 7750d5fb0eed5aadcfb300d954cd8a06c38d92001a4babb24681e14257ee1005" id=c2015e92-3d48-407a-b66b-b5478b70fe52 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:16:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:16:59.995802749Z" level=info msg="Started container" PID=6410 containerID=7750d5fb0eed5aadcfb300d954cd8a06c38d92001a4babb24681e14257ee1005 description=openshift-monitoring/alertmanager-main-0/kube-rbac-proxy id=c2015e92-3d48-407a-b66b-b5478b70fe52 name=/runtime.v1.RuntimeService/StartContainer sandboxID=0bb16724854db8df010963df02e5853c885cede9ad0447f86da449f730ebabf0 Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.015703095Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=ec1d2c36-a0c3-4639-a374-cfeaccfcd9f3 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.015963432Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=ec1d2c36-a0c3-4639-a374-cfeaccfcd9f3 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.017076712Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=9220f186-fd03-4707-af7d-09938044eb20 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.017279241Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=9220f186-fd03-4707-af7d-09938044eb20 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.018484288Z" level=info msg="Creating container: openshift-monitoring/alertmanager-main-0/kube-rbac-proxy-metric" id=c5a7772c-7d10-448b-b604-1470dc5e3218 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.018603976Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:17:00.110380 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" event=&{ID:fc45e847-8f55-4c48-8c48-4bfee2419dca Type:ContainerStarted Data:5c8fab5cf2da97490127bff84313a78ffc021feac020252239d18a75f21cbc6b} Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:17:00.113350 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" event=&{ID:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e Type:ContainerStarted Data:2dca518d6d28aeafb7a13c3ccc01e3f5af14216c4354b60ce83f720c1064b4b7} Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:17:00.116593 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event=&{ID:05c9a6ea-140a-49a2-bde5-f220b05aa252 Type:ContainerStarted Data:7750d5fb0eed5aadcfb300d954cd8a06c38d92001a4babb24681e14257ee1005} Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:17:00.116619 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event=&{ID:05c9a6ea-140a-49a2-bde5-f220b05aa252 Type:ContainerStarted Data:43228eea868ea1a6b603b81a7625cda951823b52f39f51806c1ba0ccdccbda56} Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:17:00.116634 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event=&{ID:05c9a6ea-140a-49a2-bde5-f220b05aa252 Type:ContainerStarted Data:744edbc686210b5388467d89b22a8f7cad5ce589c1225d191c8e1a4a5f5b4413} Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:17:00.116649 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event=&{ID:05c9a6ea-140a-49a2-bde5-f220b05aa252 Type:ContainerStarted Data:006114a568589213245ba8263ebfa77e665fa4cac909846666c0dede6b7512f9} Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:17:00.121081 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" event=&{ID:7c63c9be-d1ad-44b6-8b33-09af3ee314af Type:ContainerStarted Data:722bd8448df2ddebd63be975ab1961567271a1f1cca344e95f989bcc07d89d6e} Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:17:00.123524 2171 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:17:00.142153 2171 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:17:00.142187 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" event=&{ID:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 Type:ContainerStarted Data:30b2d68c19c82a725bb94bba811fe395846878cd0d20034d15397fe68fbe2865} Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:17:00.147645 2171 generic.go:296] "Generic (PLEG): container finished" podID=4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 containerID="87856f0f10585dd547492ed5711f26f3409ec28f07c808cd58bb845c8f6d7522" exitCode=0 Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:17:00.147678 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event=&{ID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Type:ContainerDied Data:87856f0f10585dd547492ed5711f26f3409ec28f07c808cd58bb845c8f6d7522} Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.148256004Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:91e7d68b28aa00e209065604b2b9ac1400671d9102581e8b49b37726acd7b6e4" id=5b09e5b1-9550-4c29-8806-1492b6ea0974 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.148505082Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b68b0fa7ea4653b56a91a8c213d456cab2f645207af609598f9877a41585eeae,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:91e7d68b28aa00e209065604b2b9ac1400671d9102581e8b49b37726acd7b6e4],Size_:493220384,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=5b09e5b1-9550-4c29-8806-1492b6ea0974 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.153086291Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:91e7d68b28aa00e209065604b2b9ac1400671d9102581e8b49b37726acd7b6e4" id=2ed8f530-6df4-4089-a3d4-98ba16645d8a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.153277178Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b68b0fa7ea4653b56a91a8c213d456cab2f645207af609598f9877a41585eeae,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:91e7d68b28aa00e209065604b2b9ac1400671d9102581e8b49b37726acd7b6e4],Size_:493220384,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=2ed8f530-6df4-4089-a3d4-98ba16645d8a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.154320554Z" level=info msg="Created container eb7fc5ba8ff324ca58d6a410bedf58010254481342f59413cbb3fa0f445a6348: openshift-monitoring/alertmanager-main-0/kube-rbac-proxy-metric" id=c5a7772c-7d10-448b-b604-1470dc5e3218 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.154536837Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-0/prometheus" id=2edade41-a037-4637-9405-9c1e2a58ab1a name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.154640033Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.156678529Z" level=info msg="Starting container: eb7fc5ba8ff324ca58d6a410bedf58010254481342f59413cbb3fa0f445a6348" id=c83cc4bc-ae67-48f0-a1bb-150eba2ad3de name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.188601064Z" level=info msg="Started container" PID=6448 containerID=eb7fc5ba8ff324ca58d6a410bedf58010254481342f59413cbb3fa0f445a6348 description=openshift-monitoring/alertmanager-main-0/kube-rbac-proxy-metric id=c83cc4bc-ae67-48f0-a1bb-150eba2ad3de name=/runtime.v1.RuntimeService/StartContainer sandboxID=0bb16724854db8df010963df02e5853c885cede9ad0447f86da449f730ebabf0 Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.249383896Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816" id=e1e961c8-b5eb-4581-9717-5cb7769bb850 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.249676354Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:955ee2b41ecd88e3a9c0183a0b65d6edbf524cefa1318837dcbc82b0466c4d55,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816],Size_:309576938,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=e1e961c8-b5eb-4581-9717-5cb7769bb850 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.252584429Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816" id=75cd6402-884b-4e0a-948f-31962643d01a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.252811332Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:955ee2b41ecd88e3a9c0183a0b65d6edbf524cefa1318837dcbc82b0466c4d55,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816],Size_:309576938,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=75cd6402-884b-4e0a-948f-31962643d01a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.267620185Z" level=info msg="Creating container: openshift-monitoring/alertmanager-main-0/prom-label-proxy" id=cac39f88-ead6-408c-9603-53946dbb0196 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.267815903Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.456179936Z" level=info msg="Created container 46be80933179162fd6984139645807e25e844c3795d674a8b04f734457efcc64: openshift-monitoring/prometheus-k8s-0/prometheus" id=2edade41-a037-4637-9405-9c1e2a58ab1a name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.456867947Z" level=info msg="Starting container: 46be80933179162fd6984139645807e25e844c3795d674a8b04f734457efcc64" id=1219036a-4ebb-4fc7-a2c6-7764e3a47045 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.479053828Z" level=info msg="Started container" PID=6505 containerID=46be80933179162fd6984139645807e25e844c3795d674a8b04f734457efcc64 description=openshift-monitoring/prometheus-k8s-0/prometheus id=1219036a-4ebb-4fc7-a2c6-7764e3a47045 name=/runtime.v1.RuntimeService/StartContainer sandboxID=33e5ecc33cac76d9879725e41d559c60e7711f7e5422fe607edbdadb5460eb11 Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.494630755Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=46cd5589-f11d-4568-95bf-8786a8bb0acc name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.494858697Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=46cd5589-f11d-4568-95bf-8786a8bb0acc name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.495966674Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=8b248db7-9425-4c57-bf1d-93552e73640e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.496167751Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=8b248db7-9425-4c57-bf1d-93552e73640e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.496292893Z" level=info msg="Created container 136db9dd8f79fe31a8bcb24721d70a69673af9677342ce64467024c361304f6e: openshift-monitoring/alertmanager-main-0/prom-label-proxy" id=cac39f88-ead6-408c-9603-53946dbb0196 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.497664756Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-0/config-reloader" id=e02207b5-6088-4796-90ee-d45b8a27e3f6 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.497800279Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.498991225Z" level=info msg="Starting container: 136db9dd8f79fe31a8bcb24721d70a69673af9677342ce64467024c361304f6e" id=539e1b35-116f-4582-9188-f5da813c04d3 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.542257460Z" level=info msg="Started container" PID=6535 containerID=136db9dd8f79fe31a8bcb24721d70a69673af9677342ce64467024c361304f6e description=openshift-monitoring/alertmanager-main-0/prom-label-proxy id=539e1b35-116f-4582-9188-f5da813c04d3 name=/runtime.v1.RuntimeService/StartContainer sandboxID=0bb16724854db8df010963df02e5853c885cede9ad0447f86da449f730ebabf0 Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.788378972Z" level=info msg="Created container 5456d8a84c429fbed32ddf7f23ca6a6e2863d0fc67333c5d2c2fac4f1a3329e0: openshift-monitoring/prometheus-k8s-0/config-reloader" id=e02207b5-6088-4796-90ee-d45b8a27e3f6 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.788892301Z" level=info msg="Starting container: 5456d8a84c429fbed32ddf7f23ca6a6e2863d0fc67333c5d2c2fac4f1a3329e0" id=1fb7a63a-1148-4688-b5c6-142014e1a858 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.806546250Z" level=info msg="Started container" PID=6629 containerID=5456d8a84c429fbed32ddf7f23ca6a6e2863d0fc67333c5d2c2fac4f1a3329e0 description=openshift-monitoring/prometheus-k8s-0/config-reloader id=1fb7a63a-1148-4688-b5c6-142014e1a858 name=/runtime.v1.RuntimeService/StartContainer sandboxID=33e5ecc33cac76d9879725e41d559c60e7711f7e5422fe607edbdadb5460eb11 Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.823440996Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc" id=3960bae7-b046-4934-a29e-1effbaa1d8be name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.823642708Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:08ff7a43d9185529520a29e114f0c28b070000b464a44b347240015f3b466488,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc],Size_:368278089,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=3960bae7-b046-4934-a29e-1effbaa1d8be name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.824354617Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc" id=2d0fd542-a612-4549-90e3-474695477c98 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.824542248Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:08ff7a43d9185529520a29e114f0c28b070000b464a44b347240015f3b466488,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc],Size_:368278089,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=2d0fd542-a612-4549-90e3-474695477c98 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.825293143Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-0/thanos-sidecar" id=4b84173f-c130-4144-aacd-b6ae1b56d3d9 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:17:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:00.825428194Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:17:01.156041 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event=&{ID:05c9a6ea-140a-49a2-bde5-f220b05aa252 Type:ContainerStarted Data:136db9dd8f79fe31a8bcb24721d70a69673af9677342ce64467024c361304f6e} Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:17:01.156081 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event=&{ID:05c9a6ea-140a-49a2-bde5-f220b05aa252 Type:ContainerStarted Data:eb7fc5ba8ff324ca58d6a410bedf58010254481342f59413cbb3fa0f445a6348} Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:17:01.166839 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event=&{ID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Type:ContainerStarted Data:5456d8a84c429fbed32ddf7f23ca6a6e2863d0fc67333c5d2c2fac4f1a3329e0} Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:17:01.166874 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event=&{ID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Type:ContainerStarted Data:46be80933179162fd6984139645807e25e844c3795d674a8b04f734457efcc64} Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:01.171322058Z" level=info msg="Created container ce5c1f9ab4c7e7c8e14b69367c484d06af071c61ef13ea7733af64251d5a919e: openshift-monitoring/prometheus-k8s-0/thanos-sidecar" id=4b84173f-c130-4144-aacd-b6ae1b56d3d9 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:01.171703035Z" level=info msg="Starting container: ce5c1f9ab4c7e7c8e14b69367c484d06af071c61ef13ea7733af64251d5a919e" id=ac713b95-c145-40e0-824c-c6a43766836f name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:01.292895530Z" level=info msg="Started container" PID=6676 containerID=ce5c1f9ab4c7e7c8e14b69367c484d06af071c61ef13ea7733af64251d5a919e description=openshift-monitoring/prometheus-k8s-0/thanos-sidecar id=ac713b95-c145-40e0-824c-c6a43766836f name=/runtime.v1.RuntimeService/StartContainer sandboxID=33e5ecc33cac76d9879725e41d559c60e7711f7e5422fe607edbdadb5460eb11 Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:01.325061379Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=83fd1d81-0bbf-496f-af83-8fb2e9afad9c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:01.325294370Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=83fd1d81-0bbf-496f-af83-8fb2e9afad9c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:01.326244576Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=ac47a3d1-17f3-475c-988c-93545622e9c8 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:01.326457507Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=ac47a3d1-17f3-475c-988c-93545622e9c8 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:01.327600263Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-0/prometheus-proxy" id=eeb8a12e-f703-4a23-a68c-49ecb96b68fd name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:01.327709372Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:01.646101827Z" level=info msg="Created container bab9c13d5ff60b04f64724b0be19b71d020c39b00bbce0a42b84b6baf9cf1858: openshift-monitoring/prometheus-k8s-0/prometheus-proxy" id=eeb8a12e-f703-4a23-a68c-49ecb96b68fd name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:01.646543309Z" level=info msg="Starting container: bab9c13d5ff60b04f64724b0be19b71d020c39b00bbce0a42b84b6baf9cf1858" id=0c9d5acb-621c-410e-a2e6-b5834bf2fd4f name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:01.654514876Z" level=info msg="Started container" PID=6752 containerID=bab9c13d5ff60b04f64724b0be19b71d020c39b00bbce0a42b84b6baf9cf1858 description=openshift-monitoring/prometheus-k8s-0/prometheus-proxy id=0c9d5acb-621c-410e-a2e6-b5834bf2fd4f name=/runtime.v1.RuntimeService/StartContainer sandboxID=33e5ecc33cac76d9879725e41d559c60e7711f7e5422fe607edbdadb5460eb11 Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:01.665871016Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=ea29b2e4-f254-47f6-9ccf-dc7a4dc7dba3 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:01.666098718Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=ea29b2e4-f254-47f6-9ccf-dc7a4dc7dba3 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:01.667090780Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=f806662f-3ee6-4f44-9e6d-dabab5369f79 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:01.667270998Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=f806662f-3ee6-4f44-9e6d-dabab5369f79 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:01.668458664Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-0/kube-rbac-proxy" id=1f69549d-e655-451e-8f34-dc5a16a1bd82 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:01.668580781Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:01.807199621Z" level=info msg="Created container af0c9f026ecfe440a5fc88294737d82bee4397499028478469868d8b733da42a: openshift-monitoring/prometheus-k8s-0/kube-rbac-proxy" id=1f69549d-e655-451e-8f34-dc5a16a1bd82 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:01.807717764Z" level=info msg="Starting container: af0c9f026ecfe440a5fc88294737d82bee4397499028478469868d8b733da42a" id=acd84c62-60f4-4781-9fbd-3272ba4f125c name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:01.823280700Z" level=info msg="Started container" PID=6790 containerID=af0c9f026ecfe440a5fc88294737d82bee4397499028478469868d8b733da42a description=openshift-monitoring/prometheus-k8s-0/kube-rbac-proxy id=acd84c62-60f4-4781-9fbd-3272ba4f125c name=/runtime.v1.RuntimeService/StartContainer sandboxID=33e5ecc33cac76d9879725e41d559c60e7711f7e5422fe607edbdadb5460eb11 Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:01.835463703Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=acd19c6a-7cd3-46cd-a519-3c3fb2aebf7a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:01.835686905Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=acd19c6a-7cd3-46cd-a519-3c3fb2aebf7a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:01.836648143Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=ccf05120-9696-4aee-9817-147b7bf3acf3 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:01.836865145Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=ccf05120-9696-4aee-9817-147b7bf3acf3 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:01.838036716Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-0/kube-rbac-proxy-thanos" id=73ec2ba2-2e91-417f-834c-b88d15acba77 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:01.838149908Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:01.952130125Z" level=info msg="Created container 0bc32a41872e3fdd7958d7634b2f0a07b2c149a1e7107267df034099986b45b2: openshift-monitoring/prometheus-k8s-0/kube-rbac-proxy-thanos" id=73ec2ba2-2e91-417f-834c-b88d15acba77 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:01.952562400Z" level=info msg="Starting container: 0bc32a41872e3fdd7958d7634b2f0a07b2c149a1e7107267df034099986b45b2" id=1b250656-d7ce-4a48-b837-7b81e8399ed8 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:17:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:01.961204027Z" level=info msg="Started container" PID=6827 containerID=0bc32a41872e3fdd7958d7634b2f0a07b2c149a1e7107267df034099986b45b2 description=openshift-monitoring/prometheus-k8s-0/kube-rbac-proxy-thanos id=1b250656-d7ce-4a48-b837-7b81e8399ed8 name=/runtime.v1.RuntimeService/StartContainer sandboxID=33e5ecc33cac76d9879725e41d559c60e7711f7e5422fe607edbdadb5460eb11 Jan 05 09:17:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:17:02.164336 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event=&{ID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Type:ContainerStarted Data:0bc32a41872e3fdd7958d7634b2f0a07b2c149a1e7107267df034099986b45b2} Jan 05 09:17:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:17:02.164370 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event=&{ID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Type:ContainerStarted Data:af0c9f026ecfe440a5fc88294737d82bee4397499028478469868d8b733da42a} Jan 05 09:17:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:17:02.164386 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event=&{ID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Type:ContainerStarted Data:bab9c13d5ff60b04f64724b0be19b71d020c39b00bbce0a42b84b6baf9cf1858} Jan 05 09:17:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:17:02.164440 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event=&{ID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Type:ContainerStarted Data:ce5c1f9ab4c7e7c8e14b69367c484d06af071c61ef13ea7733af64251d5a919e} Jan 05 09:17:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:17:04.169547 2171 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 09:17:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:17:04.580707 2171 kubelet.go:2229] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:17:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:17:04.763854 2171 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:17:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:06.479133142Z" level=warning msg="Found defunct process with PID 6300 (haproxy)" Jan 05 09:17:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:17:09.761867 2171 kubelet.go:2229] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:17:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:17:09.807297 2171 kubelet.go:2229] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:17:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:17:10.226915 2171 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:17:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: E0105 09:17:22.640804 2171 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"302e2f3529c590dfc8f53fc352ffb3636352b16ca70ce41c7c522e56153ffe51\": container with ID starting with 302e2f3529c590dfc8f53fc352ffb3636352b16ca70ce41c7c522e56153ffe51 not found: ID does not exist" containerID="302e2f3529c590dfc8f53fc352ffb3636352b16ca70ce41c7c522e56153ffe51" Jan 05 09:17:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:17:22.640854 2171 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="302e2f3529c590dfc8f53fc352ffb3636352b16ca70ce41c7c522e56153ffe51" err="rpc error: code = NotFound desc = could not find container \"302e2f3529c590dfc8f53fc352ffb3636352b16ca70ce41c7c522e56153ffe51\": container with ID starting with 302e2f3529c590dfc8f53fc352ffb3636352b16ca70ce41c7c522e56153ffe51 not found: ID does not exist" Jan 05 09:17:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:17:24.626568 2171 kubelet.go:2229] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:17:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:17:25.533777 2171 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:17:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:27.633223392Z" level=warning msg="Found defunct process with PID 7015 (haproxy)" Jan 05 09:17:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:17:36.479369211Z" level=warning msg="Found defunct process with PID 7015 (haproxy)" Jan 05 09:18:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:18:27.635392856Z" level=warning msg="Found defunct process with PID 9013 (haproxy)" Jan 05 09:18:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:18:36.492126753Z" level=warning msg="Found defunct process with PID 9013 (haproxy)" Jan 05 09:18:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:18:36.492224325Z" level=warning msg="Found defunct process with PID 9233 (haproxy)" Jan 05 09:18:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:18:36.492289513Z" level=warning msg="Found defunct process with PID 9508 (haproxy)" Jan 05 09:18:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:18:52.309973 2171 logs.go:405] "Finished parsing log file, hit bytes limit" path="/var/log/pods/openshift-sdn_sdn-fbccx_b5d621a5-07d9-459b-8745-a991cf4a179e/sdn/1.log" limit=65536 Jan 05 09:18:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:18:52.359634 2171 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-sdn_sdn-fbccx_b5d621a5-07d9-459b-8745-a991cf4a179e/kube-rbac-proxy/1.log" Jan 05 09:18:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:18:57.635447576Z" level=warning msg="Found defunct process with PID 7508 (haproxy)" Jan 05 09:18:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:18:57.635580320Z" level=warning msg="Found defunct process with PID 10293 (haproxy)" Jan 05 09:18:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:18:57.635647412Z" level=warning msg="Found defunct process with PID 10699 (iptables)" Jan 05 09:19:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:19:06.482773915Z" level=warning msg="Found defunct process with PID 7508 (haproxy)" Jan 05 09:21:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:21:22.519853 2171 kubelet.go:1343] "Image garbage collection succeeded" Jan 05 09:21:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:21:22.735842122Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=1a1d522d-b342-4d46-a3a9-47625dfbd04c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:21:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:21:22.736078588Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=1a1d522d-b342-4d46-a3a9-47625dfbd04c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:21:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:21:36.479160400Z" level=warning msg="Found defunct process with PID 9973 (haproxy)" Jan 05 09:22:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:04.775709283Z" level=warning msg="Stdout copy error: read /dev/ptmx: input/output error" Jan 05 09:22:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:06.480302157Z" level=warning msg="Found defunct process with PID 10529 (haproxy)" Jan 05 09:22:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:18.341153 2171 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-sdn/sdn-fbccx] Jan 05 09:22:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:18.341664 2171 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-sdn/sdn-fbccx" podUID=b5d621a5-07d9-459b-8745-a991cf4a179e containerName="kube-rbac-proxy" containerID="cri-o://b84dab7a298b3be0b8ad3e96ce486984126dd9280a93922e23840d85810b4733" gracePeriod=30 Jan 05 09:22:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:18.346100300Z" level=info msg="Stopping container: b84dab7a298b3be0b8ad3e96ce486984126dd9280a93922e23840d85810b4733 (timeout: 30s)" id=5f88ea53-26a6-4bdf-a3b1-6ec59f8d24d5 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:22:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:18.428892358Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 09:22:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:18.436690 2171 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-sdn/sdn-fbccx" podUID=b5d621a5-07d9-459b-8745-a991cf4a179e containerName="sdn" containerID="cri-o://13ea8c9b5aa383f0369dbc0f8fb24015e3ec8833108b7bfb08614c7e60ccc20b" gracePeriod=30 Jan 05 09:22:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:18.437129257Z" level=info msg="Stopping container: 13ea8c9b5aa383f0369dbc0f8fb24015e3ec8833108b7bfb08614c7e60ccc20b (timeout: 30s)" id=a5e7cbcc-3d14-47da-bf36-a32a0a9b1554 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:22:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:18.697020 2171 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-multus/multus-twl8f] Jan 05 09:22:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:18.697224 2171 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-multus/multus-twl8f" podUID=e13f5318-440f-4b4e-863d-2489155b9feb containerName="kube-multus" containerID="cri-o://e6c544472a3fe1c13d8c25f801c04018deedf4070a00354504f4f053617bcbec" gracePeriod=10 Jan 05 09:22:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:18.698059801Z" level=info msg="Stopping container: e6c544472a3fe1c13d8c25f801c04018deedf4070a00354504f4f053617bcbec (timeout: 10s)" id=d4dad94c-fe92-4f54-870d-5d1cc07e2214 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:22:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:18.978917930Z" level=info msg="Stopped container e6c544472a3fe1c13d8c25f801c04018deedf4070a00354504f4f053617bcbec: openshift-multus/multus-twl8f/kube-multus" id=d4dad94c-fe92-4f54-870d-5d1cc07e2214 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:22:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:18.979511006Z" level=info msg="Stopping pod sandbox: d04b2e4ec3af58e65ad07d97f2565affd7e50b26af21f25245500b2278268045" id=fe0b1684-e150-4c15-8a6b-f04ffeccd278 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:19.053808159Z" level=info msg="Stopped pod sandbox: d04b2e4ec3af58e65ad07d97f2565affd7e50b26af21f25245500b2278268045" id=fe0b1684-e150-4c15-8a6b-f04ffeccd278 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.085784 2171 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-os-release\") pod \"e13f5318-440f-4b4e-863d-2489155b9feb\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.085856 2171 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9799x\" (UniqueName: \"kubernetes.io/projected/e13f5318-440f-4b4e-863d-2489155b9feb-kube-api-access-9799x\") pod \"e13f5318-440f-4b4e-863d-2489155b9feb\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.085884 2171 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-multus-cni-dir\") pod \"e13f5318-440f-4b4e-863d-2489155b9feb\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.085924 2171 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-cnibin\") pod \"e13f5318-440f-4b4e-863d-2489155b9feb\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.085954 2171 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e13f5318-440f-4b4e-863d-2489155b9feb-cni-binary-copy\") pod \"e13f5318-440f-4b4e-863d-2489155b9feb\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.085981 2171 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-system-cni-dir\") pod \"e13f5318-440f-4b4e-863d-2489155b9feb\" (UID: \"e13f5318-440f-4b4e-863d-2489155b9feb\") " Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.086153 2171 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-system-cni-dir" (OuterVolumeSpecName: "system-cni-dir") pod "e13f5318-440f-4b4e-863d-2489155b9feb" (UID: "e13f5318-440f-4b4e-863d-2489155b9feb"). InnerVolumeSpecName "system-cni-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.086197 2171 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-os-release" (OuterVolumeSpecName: "os-release") pod "e13f5318-440f-4b4e-863d-2489155b9feb" (UID: "e13f5318-440f-4b4e-863d-2489155b9feb"). InnerVolumeSpecName "os-release". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.086524 2171 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-cnibin" (OuterVolumeSpecName: "cnibin") pod "e13f5318-440f-4b4e-863d-2489155b9feb" (UID: "e13f5318-440f-4b4e-863d-2489155b9feb"). InnerVolumeSpecName "cnibin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.086565 2171 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-multus-cni-dir" (OuterVolumeSpecName: "multus-cni-dir") pod "e13f5318-440f-4b4e-863d-2489155b9feb" (UID: "e13f5318-440f-4b4e-863d-2489155b9feb"). InnerVolumeSpecName "multus-cni-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 09:22:19.086711 2171 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/e13f5318-440f-4b4e-863d-2489155b9feb/volumes/kubernetes.io~configmap/cni-binary-copy: clearQuota called, but quotas disabled Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.086915 2171 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e13f5318-440f-4b4e-863d-2489155b9feb-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "e13f5318-440f-4b4e-863d-2489155b9feb" (UID: "e13f5318-440f-4b4e-863d-2489155b9feb"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.101113 2171 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e13f5318-440f-4b4e-863d-2489155b9feb-kube-api-access-9799x" (OuterVolumeSpecName: "kube-api-access-9799x") pod "e13f5318-440f-4b4e-863d-2489155b9feb" (UID: "e13f5318-440f-4b4e-863d-2489155b9feb"). InnerVolumeSpecName "kube-api-access-9799x". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.113170 2171 generic.go:296] "Generic (PLEG): container finished" podID=b5d621a5-07d9-459b-8745-a991cf4a179e containerID="13ea8c9b5aa383f0369dbc0f8fb24015e3ec8833108b7bfb08614c7e60ccc20b" exitCode=0 Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.113234 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-sdn/sdn-fbccx" event=&{ID:b5d621a5-07d9-459b-8745-a991cf4a179e Type:ContainerDied Data:13ea8c9b5aa383f0369dbc0f8fb24015e3ec8833108b7bfb08614c7e60ccc20b} Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.114566 2171 generic.go:296] "Generic (PLEG): container finished" podID=e13f5318-440f-4b4e-863d-2489155b9feb containerID="e6c544472a3fe1c13d8c25f801c04018deedf4070a00354504f4f053617bcbec" exitCode=0 Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.114598 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-twl8f" event=&{ID:e13f5318-440f-4b4e-863d-2489155b9feb Type:ContainerDied Data:e6c544472a3fe1c13d8c25f801c04018deedf4070a00354504f4f053617bcbec} Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.114620 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-twl8f" event=&{ID:e13f5318-440f-4b4e-863d-2489155b9feb Type:ContainerDied Data:d04b2e4ec3af58e65ad07d97f2565affd7e50b26af21f25245500b2278268045} Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.114655 2171 scope.go:115] "RemoveContainer" containerID="e6c544472a3fe1c13d8c25f801c04018deedf4070a00354504f4f053617bcbec" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:19.125858039Z" level=info msg="Removing container: e6c544472a3fe1c13d8c25f801c04018deedf4070a00354504f4f053617bcbec" id=8d7236c2-37a9-4704-8326-db4b9852146b name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:19.137968201Z" level=info msg="Stopped container 13ea8c9b5aa383f0369dbc0f8fb24015e3ec8833108b7bfb08614c7e60ccc20b: openshift-sdn/sdn-fbccx/sdn" id=a5e7cbcc-3d14-47da-bf36-a32a0a9b1554 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.149359 2171 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-multus/multus-twl8f] Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.155052 2171 kubelet.go:2129] "SyncLoop REMOVE" source="api" pods=[openshift-multus/multus-twl8f] Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:19.156529488Z" level=info msg="Removed container e6c544472a3fe1c13d8c25f801c04018deedf4070a00354504f4f053617bcbec: openshift-multus/multus-twl8f/kube-multus" id=8d7236c2-37a9-4704-8326-db4b9852146b name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.156815 2171 scope.go:115] "RemoveContainer" containerID="e6c544472a3fe1c13d8c25f801c04018deedf4070a00354504f4f053617bcbec" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: E0105 09:22:19.157100 2171 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6c544472a3fe1c13d8c25f801c04018deedf4070a00354504f4f053617bcbec\": container with ID starting with e6c544472a3fe1c13d8c25f801c04018deedf4070a00354504f4f053617bcbec not found: ID does not exist" containerID="e6c544472a3fe1c13d8c25f801c04018deedf4070a00354504f4f053617bcbec" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.157137 2171 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:e6c544472a3fe1c13d8c25f801c04018deedf4070a00354504f4f053617bcbec} err="failed to get container status \"e6c544472a3fe1c13d8c25f801c04018deedf4070a00354504f4f053617bcbec\": rpc error: code = NotFound desc = could not find container \"e6c544472a3fe1c13d8c25f801c04018deedf4070a00354504f4f053617bcbec\": container with ID starting with e6c544472a3fe1c13d8c25f801c04018deedf4070a00354504f4f053617bcbec not found: ID does not exist" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.167706 2171 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-multus/multus-pbfzz] Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.167759 2171 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: E0105 09:22:19.167848 2171 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="e13f5318-440f-4b4e-863d-2489155b9feb" containerName="kube-multus" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.167864 2171 state_mem.go:107] "Deleted CPUSet assignment" podUID="e13f5318-440f-4b4e-863d-2489155b9feb" containerName="kube-multus" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.167931 2171 memory_manager.go:345] "RemoveStaleState removing state" podUID="e13f5318-440f-4b4e-863d-2489155b9feb" containerName="kube-multus" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.187064 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tv6rn\" (UniqueName: \"kubernetes.io/projected/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-kube-api-access-tv6rn\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.187107 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-cnibin\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.187144 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-cni-binary-copy\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.187177 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-system-cni-dir\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.187204 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-multus-cni-dir\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.187237 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-os-release\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.187300 2171 reconciler.go:399] "Volume detached for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-cnibin\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.187320 2171 reconciler.go:399] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e13f5318-440f-4b4e-863d-2489155b9feb-cni-binary-copy\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.187340 2171 reconciler.go:399] "Volume detached for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-system-cni-dir\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.187357 2171 reconciler.go:399] "Volume detached for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-os-release\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.187374 2171 reconciler.go:399] "Volume detached for volume \"kube-api-access-9799x\" (UniqueName: \"kubernetes.io/projected/e13f5318-440f-4b4e-863d-2489155b9feb-kube-api-access-9799x\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.187391 2171 reconciler.go:399] "Volume detached for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e13f5318-440f-4b4e-863d-2489155b9feb-multus-cni-dir\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.288620 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-cnibin\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.288672 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-cni-binary-copy\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.288722 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-system-cni-dir\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.288762 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-multus-cni-dir\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.288798 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-os-release\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.288847 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-tv6rn\" (UniqueName: \"kubernetes.io/projected/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-kube-api-access-tv6rn\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.289061 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-system-cni-dir\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.289130 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-cnibin\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.289224 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-multus-cni-dir\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.289319 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-os-release\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.289641 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-cni-binary-copy\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.306669 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-tv6rn\" (UniqueName: \"kubernetes.io/projected/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-kube-api-access-tv6rn\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.491892 2171 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-pbfzz" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:19.493565104Z" level=info msg="Running pod sandbox: openshift-multus/multus-pbfzz/POD" id=855ed793-059f-4306-b602-ce321911149d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:19.493618845Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:19.545703910Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=855ed793-059f-4306-b602-ce321911149d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:19.567011084Z" level=info msg="Stopped container b84dab7a298b3be0b8ad3e96ce486984126dd9280a93922e23840d85810b4733: openshift-sdn/sdn-fbccx/kube-rbac-proxy" id=5f88ea53-26a6-4bdf-a3b1-6ec59f8d24d5 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:19.567490846Z" level=info msg="Stopping pod sandbox: 66a629012ef7e22f6af44b09d2cda79cab33d282fd659514ee53da6e316c1302" id=c3ad7b23-8f81-4818-ac80-c1e05a4d154b name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 09:22:19.567274 2171 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf907add9_a2cf_4fc7_b7f8_f283bc1c2dd7.slice/crio-7037be6a4b21795a75245d556479196ac813418de71922918e5a443959aa8e0e.scope WatchSource:0}: Error finding container 7037be6a4b21795a75245d556479196ac813418de71922918e5a443959aa8e0e: Status 404 returned error can't find the container with id 7037be6a4b21795a75245d556479196ac813418de71922918e5a443959aa8e0e Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:19.573599745Z" level=info msg="Ran pod sandbox 7037be6a4b21795a75245d556479196ac813418de71922918e5a443959aa8e0e with infra container: openshift-multus/multus-pbfzz/POD" id=855ed793-059f-4306-b602-ce321911149d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:19.574411200Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=fdf6190c-fcef-4260-818a-71556691fdd1 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:19.574608277Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=fdf6190c-fcef-4260-818a-71556691fdd1 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:19.575313311Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=9c792c58-57c0-489b-adfb-42034de1c093 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:19.575490856Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=9c792c58-57c0-489b-adfb-42034de1c093 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:19.576513401Z" level=info msg="Creating container: openshift-multus/multus-pbfzz/kube-multus" id=734eebc2-b090-4448-9cab-bf2e8b52045c name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:19.576580242Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:19.612793383Z" level=info msg="Stopped pod sandbox: 66a629012ef7e22f6af44b09d2cda79cab33d282fd659514ee53da6e316c1302" id=c3ad7b23-8f81-4818-ac80-c1e05a4d154b name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.693158 2171 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-lib-cni-networks-openshift-sdn\") pod \"b5d621a5-07d9-459b-8745-a991cf4a179e\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.693228 2171 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b5d621a5-07d9-459b-8745-a991cf4a179e-env-overrides\") pod \"b5d621a5-07d9-459b-8745-a991cf4a179e\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.693284 2171 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"sdn-metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5d621a5-07d9-459b-8745-a991cf4a179e-sdn-metrics-certs\") pod \"b5d621a5-07d9-459b-8745-a991cf4a179e\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.693318 2171 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-slash\") pod \"b5d621a5-07d9-459b-8745-a991cf4a179e\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.693356 2171 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5d621a5-07d9-459b-8745-a991cf4a179e-config\") pod \"b5d621a5-07d9-459b-8745-a991cf4a179e\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.693401 2171 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-cni-bin\") pod \"b5d621a5-07d9-459b-8745-a991cf4a179e\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.693433 2171 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"host-cni-conf\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-cni-conf\") pod \"b5d621a5-07d9-459b-8745-a991cf4a179e\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.693465 2171 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-run-netns\") pod \"b5d621a5-07d9-459b-8745-a991cf4a179e\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.693500 2171 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"host-var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-dbus\") pod \"b5d621a5-07d9-459b-8745-a991cf4a179e\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.693552 2171 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"host-var-run-ovs\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-ovs\") pod \"b5d621a5-07d9-459b-8745-a991cf4a179e\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.693589 2171 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"host-var-run-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-openshift-sdn\") pod \"b5d621a5-07d9-459b-8745-a991cf4a179e\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.693624 2171 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"host-var-run-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-kubernetes\") pod \"b5d621a5-07d9-459b-8745-a991cf4a179e\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.693663 2171 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"config-mtu-migration\" (UniqueName: \"kubernetes.io/configmap/b5d621a5-07d9-459b-8745-a991cf4a179e-config-mtu-migration\") pod \"b5d621a5-07d9-459b-8745-a991cf4a179e\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.693688 2171 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"host-modules\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-modules\") pod \"b5d621a5-07d9-459b-8745-a991cf4a179e\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.693713 2171 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-etc-sysconfig\") pod \"b5d621a5-07d9-459b-8745-a991cf4a179e\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.693744 2171 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7d6lx\" (UniqueName: \"kubernetes.io/projected/b5d621a5-07d9-459b-8745-a991cf4a179e-kube-api-access-7d6lx\") pod \"b5d621a5-07d9-459b-8745-a991cf4a179e\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.693794 2171 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"host-var-run-crio\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-crio\") pod \"b5d621a5-07d9-459b-8745-a991cf4a179e\" (UID: \"b5d621a5-07d9-459b-8745-a991cf4a179e\") " Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.693997 2171 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-crio" (OuterVolumeSpecName: "host-var-run-crio") pod "b5d621a5-07d9-459b-8745-a991cf4a179e" (UID: "b5d621a5-07d9-459b-8745-a991cf4a179e"). InnerVolumeSpecName "host-var-run-crio". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.694033 2171 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-lib-cni-networks-openshift-sdn" (OuterVolumeSpecName: "host-var-lib-cni-networks-openshift-sdn") pod "b5d621a5-07d9-459b-8745-a991cf4a179e" (UID: "b5d621a5-07d9-459b-8745-a991cf4a179e"). InnerVolumeSpecName "host-var-lib-cni-networks-openshift-sdn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 09:22:19.694219 2171 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/b5d621a5-07d9-459b-8745-a991cf4a179e/volumes/kubernetes.io~configmap/env-overrides: clearQuota called, but quotas disabled Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.694394 2171 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5d621a5-07d9-459b-8745-a991cf4a179e-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "b5d621a5-07d9-459b-8745-a991cf4a179e" (UID: "b5d621a5-07d9-459b-8745-a991cf4a179e"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.694606 2171 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-ovs" (OuterVolumeSpecName: "host-var-run-ovs") pod "b5d621a5-07d9-459b-8745-a991cf4a179e" (UID: "b5d621a5-07d9-459b-8745-a991cf4a179e"). InnerVolumeSpecName "host-var-run-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.694645 2171 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-slash" (OuterVolumeSpecName: "host-slash") pod "b5d621a5-07d9-459b-8745-a991cf4a179e" (UID: "b5d621a5-07d9-459b-8745-a991cf4a179e"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 09:22:19.694799 2171 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/b5d621a5-07d9-459b-8745-a991cf4a179e/volumes/kubernetes.io~configmap/config: clearQuota called, but quotas disabled Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.694962 2171 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5d621a5-07d9-459b-8745-a991cf4a179e-config" (OuterVolumeSpecName: "config") pod "b5d621a5-07d9-459b-8745-a991cf4a179e" (UID: "b5d621a5-07d9-459b-8745-a991cf4a179e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.694984 2171 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "b5d621a5-07d9-459b-8745-a991cf4a179e" (UID: "b5d621a5-07d9-459b-8745-a991cf4a179e"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.694999 2171 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-cni-conf" (OuterVolumeSpecName: "host-cni-conf") pod "b5d621a5-07d9-459b-8745-a991cf4a179e" (UID: "b5d621a5-07d9-459b-8745-a991cf4a179e"). InnerVolumeSpecName "host-cni-conf". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.695019 2171 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "b5d621a5-07d9-459b-8745-a991cf4a179e" (UID: "b5d621a5-07d9-459b-8745-a991cf4a179e"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.695036 2171 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-dbus" (OuterVolumeSpecName: "host-var-run-dbus") pod "b5d621a5-07d9-459b-8745-a991cf4a179e" (UID: "b5d621a5-07d9-459b-8745-a991cf4a179e"). InnerVolumeSpecName "host-var-run-dbus". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 09:22:19.695134 2171 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/b5d621a5-07d9-459b-8745-a991cf4a179e/volumes/kubernetes.io~configmap/config-mtu-migration: clearQuota called, but quotas disabled Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.695232 2171 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5d621a5-07d9-459b-8745-a991cf4a179e-config-mtu-migration" (OuterVolumeSpecName: "config-mtu-migration") pod "b5d621a5-07d9-459b-8745-a991cf4a179e" (UID: "b5d621a5-07d9-459b-8745-a991cf4a179e"). InnerVolumeSpecName "config-mtu-migration". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.695271 2171 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-openshift-sdn" (OuterVolumeSpecName: "host-var-run-openshift-sdn") pod "b5d621a5-07d9-459b-8745-a991cf4a179e" (UID: "b5d621a5-07d9-459b-8745-a991cf4a179e"). InnerVolumeSpecName "host-var-run-openshift-sdn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.695287 2171 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-kubernetes" (OuterVolumeSpecName: "host-var-run-kubernetes") pod "b5d621a5-07d9-459b-8745-a991cf4a179e" (UID: "b5d621a5-07d9-459b-8745-a991cf4a179e"). InnerVolumeSpecName "host-var-run-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.695302 2171 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-etc-sysconfig" (OuterVolumeSpecName: "etc-sysconfig") pod "b5d621a5-07d9-459b-8745-a991cf4a179e" (UID: "b5d621a5-07d9-459b-8745-a991cf4a179e"). InnerVolumeSpecName "etc-sysconfig". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.695315 2171 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-modules" (OuterVolumeSpecName: "host-modules") pod "b5d621a5-07d9-459b-8745-a991cf4a179e" (UID: "b5d621a5-07d9-459b-8745-a991cf4a179e"). InnerVolumeSpecName "host-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.706760 2171 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5d621a5-07d9-459b-8745-a991cf4a179e-sdn-metrics-certs" (OuterVolumeSpecName: "sdn-metrics-certs") pod "b5d621a5-07d9-459b-8745-a991cf4a179e" (UID: "b5d621a5-07d9-459b-8745-a991cf4a179e"). InnerVolumeSpecName "sdn-metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.706895 2171 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5d621a5-07d9-459b-8745-a991cf4a179e-kube-api-access-7d6lx" (OuterVolumeSpecName: "kube-api-access-7d6lx") pod "b5d621a5-07d9-459b-8745-a991cf4a179e" (UID: "b5d621a5-07d9-459b-8745-a991cf4a179e"). InnerVolumeSpecName "kube-api-access-7d6lx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:19.746635001Z" level=info msg="Created container 4fed50c900e7456872d20306aa0e63b13927facd5fb4913292c3036a63611a79: openshift-multus/multus-pbfzz/kube-multus" id=734eebc2-b090-4448-9cab-bf2e8b52045c name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:19.747182780Z" level=info msg="Starting container: 4fed50c900e7456872d20306aa0e63b13927facd5fb4913292c3036a63611a79" id=1894899f-3c6d-47f7-98b5-e4ab5632d1c9 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:19.756729323Z" level=info msg="Started container" PID=15626 containerID=4fed50c900e7456872d20306aa0e63b13927facd5fb4913292c3036a63611a79 description=openshift-multus/multus-pbfzz/kube-multus id=1894899f-3c6d-47f7-98b5-e4ab5632d1c9 name=/runtime.v1.RuntimeService/StartContainer sandboxID=7037be6a4b21795a75245d556479196ac813418de71922918e5a443959aa8e0e Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:19.783386259Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_fcea6a0c-198c-4f71-86a3-81733e756e98\"" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.799782 2171 reconciler.go:399] "Volume detached for volume \"host-var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-dbus\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.799824 2171 reconciler.go:399] "Volume detached for volume \"host-var-run-ovs\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-ovs\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.799849 2171 reconciler.go:399] "Volume detached for volume \"host-var-run-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-openshift-sdn\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.799868 2171 reconciler.go:399] "Volume detached for volume \"host-var-run-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-kubernetes\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.799888 2171 reconciler.go:399] "Volume detached for volume \"config-mtu-migration\" (UniqueName: \"kubernetes.io/configmap/b5d621a5-07d9-459b-8745-a991cf4a179e-config-mtu-migration\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.799906 2171 reconciler.go:399] "Volume detached for volume \"host-modules\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-modules\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.799923 2171 reconciler.go:399] "Volume detached for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-etc-sysconfig\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.799942 2171 reconciler.go:399] "Volume detached for volume \"kube-api-access-7d6lx\" (UniqueName: \"kubernetes.io/projected/b5d621a5-07d9-459b-8745-a991cf4a179e-kube-api-access-7d6lx\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.799963 2171 reconciler.go:399] "Volume detached for volume \"host-var-run-crio\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-run-crio\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.799985 2171 reconciler.go:399] "Volume detached for volume \"host-var-lib-cni-networks-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-var-lib-cni-networks-openshift-sdn\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.800004 2171 reconciler.go:399] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b5d621a5-07d9-459b-8745-a991cf4a179e-env-overrides\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.800022 2171 reconciler.go:399] "Volume detached for volume \"sdn-metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5d621a5-07d9-459b-8745-a991cf4a179e-sdn-metrics-certs\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.800041 2171 reconciler.go:399] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-slash\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.800061 2171 reconciler.go:399] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5d621a5-07d9-459b-8745-a991cf4a179e-config\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.800078 2171 reconciler.go:399] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-cni-bin\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.800096 2171 reconciler.go:399] "Volume detached for volume \"host-cni-conf\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-cni-conf\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:19.800114 2171 reconciler.go:399] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b5d621a5-07d9-459b-8745-a991cf4a179e-host-run-netns\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:19.807145845Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:19.807181244Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:19.827078490Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/multus\"" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:19.871874983Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:19.871915192Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:22:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:19.871938399Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_fcea6a0c-198c-4f71-86a3-81733e756e98\"" Jan 05 09:22:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:20.125336 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-pbfzz" event=&{ID:f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7 Type:ContainerStarted Data:4fed50c900e7456872d20306aa0e63b13927facd5fb4913292c3036a63611a79} Jan 05 09:22:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:20.125375 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-pbfzz" event=&{ID:f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7 Type:ContainerStarted Data:7037be6a4b21795a75245d556479196ac813418de71922918e5a443959aa8e0e} Jan 05 09:22:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:20.127945 2171 generic.go:296] "Generic (PLEG): container finished" podID=b5d621a5-07d9-459b-8745-a991cf4a179e containerID="b84dab7a298b3be0b8ad3e96ce486984126dd9280a93922e23840d85810b4733" exitCode=0 Jan 05 09:22:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:20.127979 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-sdn/sdn-fbccx" event=&{ID:b5d621a5-07d9-459b-8745-a991cf4a179e Type:ContainerDied Data:b84dab7a298b3be0b8ad3e96ce486984126dd9280a93922e23840d85810b4733} Jan 05 09:22:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:20.128001 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-sdn/sdn-fbccx" event=&{ID:b5d621a5-07d9-459b-8745-a991cf4a179e Type:ContainerDied Data:66a629012ef7e22f6af44b09d2cda79cab33d282fd659514ee53da6e316c1302} Jan 05 09:22:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:20.128021 2171 scope.go:115] "RemoveContainer" containerID="b84dab7a298b3be0b8ad3e96ce486984126dd9280a93922e23840d85810b4733" Jan 05 09:22:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:20.132138406Z" level=info msg="Removing container: b84dab7a298b3be0b8ad3e96ce486984126dd9280a93922e23840d85810b4733" id=704c3576-f562-4c77-a747-47f99876a095 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:22:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:20.157877591Z" level=info msg="Removed container b84dab7a298b3be0b8ad3e96ce486984126dd9280a93922e23840d85810b4733: openshift-sdn/sdn-fbccx/kube-rbac-proxy" id=704c3576-f562-4c77-a747-47f99876a095 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:22:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:20.158056 2171 scope.go:115] "RemoveContainer" containerID="13ea8c9b5aa383f0369dbc0f8fb24015e3ec8833108b7bfb08614c7e60ccc20b" Jan 05 09:22:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:20.158791409Z" level=info msg="Removing container: 13ea8c9b5aa383f0369dbc0f8fb24015e3ec8833108b7bfb08614c7e60ccc20b" id=6986d4c7-f3f9-43c6-9a0d-28bf886e9e6e name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:22:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:20.178325 2171 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-sdn/sdn-fbccx] Jan 05 09:22:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:20.185477 2171 kubelet.go:2129] "SyncLoop REMOVE" source="api" pods=[openshift-sdn/sdn-fbccx] Jan 05 09:22:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:20.207897517Z" level=info msg="Removed container 13ea8c9b5aa383f0369dbc0f8fb24015e3ec8833108b7bfb08614c7e60ccc20b: openshift-sdn/sdn-fbccx/sdn" id=6986d4c7-f3f9-43c6-9a0d-28bf886e9e6e name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:22:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:20.208110 2171 scope.go:115] "RemoveContainer" containerID="b84dab7a298b3be0b8ad3e96ce486984126dd9280a93922e23840d85810b4733" Jan 05 09:22:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: E0105 09:22:20.209437 2171 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b84dab7a298b3be0b8ad3e96ce486984126dd9280a93922e23840d85810b4733\": container with ID starting with b84dab7a298b3be0b8ad3e96ce486984126dd9280a93922e23840d85810b4733 not found: ID does not exist" containerID="b84dab7a298b3be0b8ad3e96ce486984126dd9280a93922e23840d85810b4733" Jan 05 09:22:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:20.209476 2171 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:b84dab7a298b3be0b8ad3e96ce486984126dd9280a93922e23840d85810b4733} err="failed to get container status \"b84dab7a298b3be0b8ad3e96ce486984126dd9280a93922e23840d85810b4733\": rpc error: code = NotFound desc = could not find container \"b84dab7a298b3be0b8ad3e96ce486984126dd9280a93922e23840d85810b4733\": container with ID starting with b84dab7a298b3be0b8ad3e96ce486984126dd9280a93922e23840d85810b4733 not found: ID does not exist" Jan 05 09:22:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:20.209490 2171 scope.go:115] "RemoveContainer" containerID="13ea8c9b5aa383f0369dbc0f8fb24015e3ec8833108b7bfb08614c7e60ccc20b" Jan 05 09:22:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: E0105 09:22:20.209774 2171 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13ea8c9b5aa383f0369dbc0f8fb24015e3ec8833108b7bfb08614c7e60ccc20b\": container with ID starting with 13ea8c9b5aa383f0369dbc0f8fb24015e3ec8833108b7bfb08614c7e60ccc20b not found: ID does not exist" containerID="13ea8c9b5aa383f0369dbc0f8fb24015e3ec8833108b7bfb08614c7e60ccc20b" Jan 05 09:22:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:20.209805 2171 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:13ea8c9b5aa383f0369dbc0f8fb24015e3ec8833108b7bfb08614c7e60ccc20b} err="failed to get container status \"13ea8c9b5aa383f0369dbc0f8fb24015e3ec8833108b7bfb08614c7e60ccc20b\": rpc error: code = NotFound desc = could not find container \"13ea8c9b5aa383f0369dbc0f8fb24015e3ec8833108b7bfb08614c7e60ccc20b\": container with ID starting with 13ea8c9b5aa383f0369dbc0f8fb24015e3ec8833108b7bfb08614c7e60ccc20b not found: ID does not exist" Jan 05 09:22:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:20.902303 2171 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=b5d621a5-07d9-459b-8745-a991cf4a179e path="/var/lib/kubelet/pods/b5d621a5-07d9-459b-8745-a991cf4a179e/volumes" Jan 05 09:22:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:20.902999 2171 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=e13f5318-440f-4b4e-863d-2489155b9feb path="/var/lib/kubelet/pods/e13f5318-440f-4b4e-863d-2489155b9feb/volumes" Jan 05 09:22:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:22.669627133Z" level=info msg="Stopping pod sandbox: d04b2e4ec3af58e65ad07d97f2565affd7e50b26af21f25245500b2278268045" id=80669ff6-160d-40a4-8cb5-7efdab4bf2ed name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:22:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:22.669663974Z" level=info msg="Stopped pod sandbox (already stopped): d04b2e4ec3af58e65ad07d97f2565affd7e50b26af21f25245500b2278268045" id=80669ff6-160d-40a4-8cb5-7efdab4bf2ed name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:22:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:22.669975238Z" level=info msg="Removing pod sandbox: d04b2e4ec3af58e65ad07d97f2565affd7e50b26af21f25245500b2278268045" id=e86922b6-25d3-4f6f-af58-386e9b794b16 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:22:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:22.678371821Z" level=info msg="Removed pod sandbox: d04b2e4ec3af58e65ad07d97f2565affd7e50b26af21f25245500b2278268045" id=e86922b6-25d3-4f6f-af58-386e9b794b16 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:22:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:22.678773479Z" level=info msg="Stopping pod sandbox: 66a629012ef7e22f6af44b09d2cda79cab33d282fd659514ee53da6e316c1302" id=01c6783b-df3b-4498-a44e-25232219bf41 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:22:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:22.678802121Z" level=info msg="Stopped pod sandbox (already stopped): 66a629012ef7e22f6af44b09d2cda79cab33d282fd659514ee53da6e316c1302" id=01c6783b-df3b-4498-a44e-25232219bf41 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:22:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:22.679021706Z" level=info msg="Removing pod sandbox: 66a629012ef7e22f6af44b09d2cda79cab33d282fd659514ee53da6e316c1302" id=bcdc5171-3eca-41e8-8421-4e8a8510602f name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:22:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:22.686942286Z" level=info msg="Removed pod sandbox: 66a629012ef7e22f6af44b09d2cda79cab33d282fd659514ee53da6e316c1302" id=bcdc5171-3eca-41e8-8421-4e8a8510602f name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:22:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: E0105 09:22:22.688210 2171 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97ec0f7bd248355e78e3b1b09cc34d015a0c89ef5c3e56dde06e0dea3a0adf60\": container with ID starting with 97ec0f7bd248355e78e3b1b09cc34d015a0c89ef5c3e56dde06e0dea3a0adf60 not found: ID does not exist" containerID="97ec0f7bd248355e78e3b1b09cc34d015a0c89ef5c3e56dde06e0dea3a0adf60" Jan 05 09:22:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:22.688272 2171 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="97ec0f7bd248355e78e3b1b09cc34d015a0c89ef5c3e56dde06e0dea3a0adf60" err="rpc error: code = NotFound desc = could not find container \"97ec0f7bd248355e78e3b1b09cc34d015a0c89ef5c3e56dde06e0dea3a0adf60\": container with ID starting with 97ec0f7bd248355e78e3b1b09cc34d015a0c89ef5c3e56dde06e0dea3a0adf60 not found: ID does not exist" Jan 05 09:22:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: E0105 09:22:22.688801 2171 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1087241f3735617d55881840f0a85549fa47fcf3dacd47d7eb2c41c8d2d4d78\": container with ID starting with a1087241f3735617d55881840f0a85549fa47fcf3dacd47d7eb2c41c8d2d4d78 not found: ID does not exist" containerID="a1087241f3735617d55881840f0a85549fa47fcf3dacd47d7eb2c41c8d2d4d78" Jan 05 09:22:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:22.688836 2171 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="a1087241f3735617d55881840f0a85549fa47fcf3dacd47d7eb2c41c8d2d4d78" err="rpc error: code = NotFound desc = could not find container \"a1087241f3735617d55881840f0a85549fa47fcf3dacd47d7eb2c41c8d2d4d78\": container with ID starting with a1087241f3735617d55881840f0a85549fa47fcf3dacd47d7eb2c41c8d2d4d78 not found: ID does not exist" Jan 05 09:22:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: E0105 09:22:22.689091 2171 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56f8ab4efdd91320957f5f42e237237e6768ff86ea26e753e3315a2ebb62fdf0\": container with ID starting with 56f8ab4efdd91320957f5f42e237237e6768ff86ea26e753e3315a2ebb62fdf0 not found: ID does not exist" containerID="56f8ab4efdd91320957f5f42e237237e6768ff86ea26e753e3315a2ebb62fdf0" Jan 05 09:22:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:22.689120 2171 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="56f8ab4efdd91320957f5f42e237237e6768ff86ea26e753e3315a2ebb62fdf0" err="rpc error: code = NotFound desc = could not find container \"56f8ab4efdd91320957f5f42e237237e6768ff86ea26e753e3315a2ebb62fdf0\": container with ID starting with 56f8ab4efdd91320957f5f42e237237e6768ff86ea26e753e3315a2ebb62fdf0 not found: ID does not exist" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.241188 2171 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-ovn-kubernetes/ovnkube-node-rp9bd] Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.241255 2171 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: E0105 09:22:29.241341 2171 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="b5d621a5-07d9-459b-8745-a991cf4a179e" containerName="kube-rbac-proxy" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.241355 2171 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5d621a5-07d9-459b-8745-a991cf4a179e" containerName="kube-rbac-proxy" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: E0105 09:22:29.241372 2171 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="b5d621a5-07d9-459b-8745-a991cf4a179e" containerName="sdn" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.241382 2171 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5d621a5-07d9-459b-8745-a991cf4a179e" containerName="sdn" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.241450 2171 memory_manager.go:345] "RemoveStaleState removing state" podUID="b5d621a5-07d9-459b-8745-a991cf4a179e" containerName="kube-rbac-proxy" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.241463 2171 memory_manager.go:345] "RemoveStaleState removing state" podUID="b5d621a5-07d9-459b-8745-a991cf4a179e" containerName="sdn" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 09:22:29.274424 2171 reflector.go:424] object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-vmxh7": failed to list *v1.Secret: secrets "ovn-kubernetes-node-dockercfg-vmxh7" is forbidden: User "system:node:pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl' and this object Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: E0105 09:22:29.274479 2171 reflector.go:140] object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-vmxh7": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets "ovn-kubernetes-node-dockercfg-vmxh7" is forbidden: User "system:node:pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl' and this object Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 09:22:29.274450 2171 reflector.go:424] object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert": failed to list *v1.Secret: secrets "ovn-node-metrics-cert" is forbidden: User "system:node:pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl' and this object Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: E0105 09:22:29.274499 2171 reflector.go:140] object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets "ovn-node-metrics-cert" is forbidden: User "system:node:pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl' and this object Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 09:22:29.274557 2171 reflector.go:424] object-"openshift-ovn-kubernetes"/"env-overrides": failed to list *v1.ConfigMap: configmaps "env-overrides" is forbidden: User "system:node:pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" cannot list resource "configmaps" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl' and this object Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: E0105 09:22:29.274570 2171 reflector.go:140] object-"openshift-ovn-kubernetes"/"env-overrides": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps "env-overrides" is forbidden: User "system:node:pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" cannot list resource "configmaps" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl' and this object Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 09:22:29.274558 2171 reflector.go:424] object-"openshift-ovn-kubernetes"/"ovn-cert": failed to list *v1.Secret: secrets "ovn-cert" is forbidden: User "system:node:pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl' and this object Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: E0105 09:22:29.274584 2171 reflector.go:140] object-"openshift-ovn-kubernetes"/"ovn-cert": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets "ovn-cert" is forbidden: User "system:node:pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl' and this object Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 09:22:29.274595 2171 reflector.go:424] object-"openshift-ovn-kubernetes"/"ovn-ca": failed to list *v1.ConfigMap: configmaps "ovn-ca" is forbidden: User "system:node:pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" cannot list resource "configmaps" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl' and this object Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: E0105 09:22:29.274609 2171 reflector.go:140] object-"openshift-ovn-kubernetes"/"ovn-ca": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps "ovn-ca" is forbidden: User "system:node:pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" cannot list resource "configmaps" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl' and this object Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 09:22:29.274617 2171 reflector.go:424] object-"openshift-ovn-kubernetes"/"ovnkube-config": failed to list *v1.ConfigMap: configmaps "ovnkube-config" is forbidden: User "system:node:pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" cannot list resource "configmaps" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl' and this object Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: E0105 09:22:29.274628 2171 reflector.go:140] object-"openshift-ovn-kubernetes"/"ovnkube-config": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps "ovnkube-config" is forbidden: User "system:node:pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" cannot list resource "configmaps" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl' and this object Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 09:22:29.274648 2171 reflector.go:424] object-"openshift-ovn-kubernetes"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" cannot list resource "configmaps" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl' and this object Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: E0105 09:22:29.274658 2171 reflector.go:140] object-"openshift-ovn-kubernetes"/"kube-root-ca.crt": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" cannot list resource "configmaps" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl' and this object Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 09:22:29.274665 2171 reflector.go:424] object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" cannot list resource "configmaps" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl' and this object Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: E0105 09:22:29.274676 2171 reflector.go:140] object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" cannot list resource "configmaps" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl' and this object Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.405810 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-systemd-units\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.405861 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csks4\" (UniqueName: \"kubernetes.io/projected/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-kube-api-access-csks4\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.405917 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovnkube-config\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.405952 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-run-netns\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.405987 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-etc-openvswitch\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.406022 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-run-ovn-kubernetes\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.406053 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-log-socket\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.406088 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.406116 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-node-log\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.406145 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-ca\" (UniqueName: \"kubernetes.io/configmap/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovn-ca\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.406177 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-var-lib-openvswitch\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.406215 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-run-openvswitch\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.406262 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-cni-netd\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.406301 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovn-node-metrics-cert\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.406340 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-slash\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.406375 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-env-overrides\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.406423 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-run-ovn\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.406457 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-cert\" (UniqueName: \"kubernetes.io/secret/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovn-cert\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.406494 2171 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-cni-bin\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.507186 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-slash\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.507257 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-env-overrides\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.507291 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-run-ovn\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.507320 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"ovn-cert\" (UniqueName: \"kubernetes.io/secret/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovn-cert\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.507353 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-cni-bin\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.507382 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-systemd-units\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.507413 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-csks4\" (UniqueName: \"kubernetes.io/projected/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-kube-api-access-csks4\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.507428 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-run-ovn\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.507452 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovnkube-config\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.507484 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-slash\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.507487 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-run-netns\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.507521 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-etc-openvswitch\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.507552 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-run-ovn-kubernetes\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.507582 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-log-socket\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.507613 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.507642 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-node-log\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.507673 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"ovn-ca\" (UniqueName: \"kubernetes.io/configmap/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovn-ca\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.507705 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-var-lib-openvswitch\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.507739 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-run-openvswitch\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.507787 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-cni-netd\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.507827 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovn-node-metrics-cert\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.507897 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-cni-bin\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.507944 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-systemd-units\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.507984 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-run-netns\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.508022 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-node-log\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.508056 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-var-lib-openvswitch\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.508064 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-etc-openvswitch\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.508098 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-run-openvswitch\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.508106 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-run-ovn-kubernetes\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.508137 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-cni-netd\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.508147 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-log-socket\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:29.508177 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:30.148928 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovnkube-config\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:30.258642 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-env-overrides\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:30.271094 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"ovn-cert\" (UniqueName: \"kubernetes.io/secret/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovn-cert\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:30.468515 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"ovn-ca\" (UniqueName: \"kubernetes.io/configmap/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovn-ca\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: E0105 09:22:30.508678 2171 secret.go:192] Couldn't get secret openshift-ovn-kubernetes/ovn-node-metrics-cert: failed to sync secret cache: timed out waiting for the condition Jan 05 09:22:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: E0105 09:22:30.508770 2171 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovn-node-metrics-cert podName:aac1c4d4-278d-4a53-b2f7-a75ba43c2d18 nodeName:}" failed. No retries permitted until 2023-01-05 09:22:31.008738253 +0000 UTC m=+368.858595443 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "ovn-node-metrics-cert" (UniqueName: "kubernetes.io/secret/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovn-node-metrics-cert") pod "ovnkube-node-rp9bd" (UID: "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18") : failed to sync secret cache: timed out waiting for the condition Jan 05 09:22:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:30.691643 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-csks4\" (UniqueName: \"kubernetes.io/projected/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-kube-api-access-csks4\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:31.019039 2171 reconciler.go:269] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovn-node-metrics-cert\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:31.022703 2171 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovn-node-metrics-cert\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:31.071707 2171 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:22:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:31.072592310Z" level=info msg="Running pod sandbox: openshift-ovn-kubernetes/ovnkube-node-rp9bd/POD" id=4c6da783-6763-4bea-ae0e-3bbaab517f0e name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:22:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:31.072652153Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:22:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:31.107875114Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=4c6da783-6763-4bea-ae0e-3bbaab517f0e name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:22:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: W0105 09:22:31.121029 2171 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaac1c4d4_278d_4a53_b2f7_a75ba43c2d18.slice/crio-3f3423d5651958cff05297583dd5450abb9aa783ed6eab062e62ecac5d0396ad.scope WatchSource:0}: Error finding container 3f3423d5651958cff05297583dd5450abb9aa783ed6eab062e62ecac5d0396ad: Status 404 returned error can't find the container with id 3f3423d5651958cff05297583dd5450abb9aa783ed6eab062e62ecac5d0396ad Jan 05 09:22:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:31.128151136Z" level=info msg="Ran pod sandbox 3f3423d5651958cff05297583dd5450abb9aa783ed6eab062e62ecac5d0396ad with infra container: openshift-ovn-kubernetes/ovnkube-node-rp9bd/POD" id=4c6da783-6763-4bea-ae0e-3bbaab517f0e name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:22:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:31.129432548Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:96010fa3364bd5030d4445bbc80df5c8ec18d2b5579f23f71235bcda47b38ca6" id=bceaff8f-c271-4a02-86de-b779ffb5397c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:22:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:31.129618033Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:96010fa3364bd5030d4445bbc80df5c8ec18d2b5579f23f71235bcda47b38ca6 not found" id=bceaff8f-c271-4a02-86de-b779ffb5397c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:22:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:31.130071 2171 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 05 09:22:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:31.130527650Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:96010fa3364bd5030d4445bbc80df5c8ec18d2b5579f23f71235bcda47b38ca6" id=f6c8b2ef-2d50-4a1f-95fc-30d8d08ecc47 name=/runtime.v1.ImageService/PullImage Jan 05 09:22:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:31.131310800Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:96010fa3364bd5030d4445bbc80df5c8ec18d2b5579f23f71235bcda47b38ca6\"" Jan 05 09:22:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:22:31.182293 2171 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" event=&{ID:aac1c4d4-278d-4a53-b2f7-a75ba43c2d18 Type:ContainerStarted Data:3f3423d5651958cff05297583dd5450abb9aa783ed6eab062e62ecac5d0396ad} Jan 05 09:22:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:22:31.272576847Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:96010fa3364bd5030d4445bbc80df5c8ec18d2b5579f23f71235bcda47b38ca6\"" Jan 05 09:23:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:23:04.904439 2171 plugin_watcher.go:215] "Removing socket path from desired state cache" path="/var/lib/kubelet/plugins_registry/diskplugin.csi.alibabacloud.com-reg.sock" Jan 05 09:23:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Stopping Kubernetes Kubelet... Jan 05 09:23:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2171]: I0105 09:23:04.953322 2171 dynamic_cafile_content.go:171] "Shutting down controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 05 09:23:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: kubelet.service: Succeeded. Jan 05 09:23:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Stopped Kubernetes Kubelet. Jan 05 09:23:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: kubelet.service: Consumed 25.474s CPU time Jan 05 09:23:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:23:06.493361925Z" level=warning msg="Found defunct process with PID 14987 (haproxy)" Jan 05 09:23:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Stopping Container Runtime Interface for OCI (CRI-O)... Jan 05 09:23:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2112]: time="2023-01-05 09:23:51.212280914Z" level=error msg="Failed to update container state for 1d7597386b38676c039e5b6aac3f7c678b4af05333bf527578cbc167c1f3ec5c: `/usr/bin/runc --root /run/runc --systemd-cgroup state 1d7597386b38676c039e5b6aac3f7c678b4af05333bf527578cbc167c1f3ec5c` failed: : signal: terminated" Jan 05 09:23:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: crio.service: Succeeded. Jan 05 09:23:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Stopped Container Runtime Interface for OCI (CRI-O). Jan 05 09:23:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: crio.service: Consumed 28.767s CPU time -- Reboot -- Jan 05 17:24:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Starting Container Runtime Interface for OCI (CRI-O)... Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.078210913Z" level=info msg="Starting CRI-O, version: 1.25.1-5.rhaos4.12.git6005903.el8, git: unknown(clean)" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.078556796Z" level=info msg="Node configuration value for hugetlb cgroup is true" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.078570077Z" level=info msg="Node configuration value for pid cgroup is true" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.078622007Z" level=info msg="Node configuration value for memoryswap cgroup is true" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.078629881Z" level=info msg="Node configuration value for cgroup v2 is false" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.087123437Z" level=info msg="Node configuration value for systemd CollectMode is true" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.093114100Z" level=info msg="Node configuration value for systemd AllowedCPUs is true" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.096830479Z" level=info msg="Using default capabilities: CAP_CHOWN, CAP_DAC_OVERRIDE, CAP_FSETID, CAP_FOWNER, CAP_SETGID, CAP_SETUID, CAP_SETPCAP, CAP_NET_BIND_SERVICE, CAP_KILL" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.171233330Z" level=info msg="Checkpoint/restore support disabled" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.171265263Z" level=info msg="Using seccomp default profile when unspecified: true" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.171274369Z" level=info msg="Using the internal default seccomp profile" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.171281950Z" level=info msg="AppArmor is disabled by the system or at CRI-O build-time" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.171289854Z" level=info msg="No blockio config file specified, blockio not configured" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.171296920Z" level=info msg="RDT not available in the host system" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.177248574Z" level=info msg="Conmon does support the --sync option" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.177268537Z" level=info msg="Conmon does support the --log-global-size-max option" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.180902977Z" level=info msg="Conmon does support the --sync option" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.180920092Z" level=info msg="Conmon does support the --log-global-size-max option" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.288105536Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.288136658Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.341763414Z" level=warning msg="Could not restore sandbox 14626e8a3347256c5a502841426c6a0bfa7993e08a5404798831b7d49a3dddf7: failed to Statfs \"/var/run/netns/86a5c5e8-22e7-4e2d-a573-d2c4609eca57\": no such file or directory" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.356796108Z" level=warning msg="Deleting all containers under sandbox 14626e8a3347256c5a502841426c6a0bfa7993e08a5404798831b7d49a3dddf7 since it could not be restored" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.367071188Z" level=warning msg="Could not restore sandbox 0bb16724854db8df010963df02e5853c885cede9ad0447f86da449f730ebabf0: failed to Statfs \"/var/run/netns/a88b2c9c-80dc-4225-a81d-ec8a37f3cf76\": no such file or directory" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.376134477Z" level=warning msg="Deleting all containers under sandbox 0bb16724854db8df010963df02e5853c885cede9ad0447f86da449f730ebabf0 since it could not be restored" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.434712406Z" level=warning msg="Could not restore sandbox c41ae071597e07deb9c5ea067a36e06b0f2b19fc9c9d6633270660715cffd374: failed to Statfs \"/var/run/netns/ba4114e0-4acf-48ca-8e88-0e71adb9d981\": no such file or directory" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.443115680Z" level=warning msg="Deleting all containers under sandbox c41ae071597e07deb9c5ea067a36e06b0f2b19fc9c9d6633270660715cffd374 since it could not be restored" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.457140192Z" level=warning msg="Could not restore sandbox 9454780ea94304660f15d0806d1edaf679bb7ba5c8879c21dffe722e3df07991: failed to Statfs \"/var/run/netns/a955ab33-d3fa-4a06-b4e3-c37811469688\": no such file or directory" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.464068948Z" level=warning msg="Deleting all containers under sandbox 9454780ea94304660f15d0806d1edaf679bb7ba5c8879c21dffe722e3df07991 since it could not be restored" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.475221554Z" level=warning msg="Could not restore sandbox bf8032a2230aa8777b5ec76b9ff9222eb0d8d4700ec6b91316ee8cd870968d80: failed to Statfs \"/var/run/netns/a033356f-4162-40fb-ac68-ff35ebe0d82e\": no such file or directory" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.482854150Z" level=warning msg="Deleting all containers under sandbox bf8032a2230aa8777b5ec76b9ff9222eb0d8d4700ec6b91316ee8cd870968d80 since it could not be restored" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.491845146Z" level=warning msg="Could not restore sandbox 63f26fe630a86618ccb6926337ae0f77635a06eec51dd1cb5b91c86eaefd519e: failed to Statfs \"/var/run/netns/e6fe9a72-a6aa-41d5-86f7-be1f22acf516\": no such file or directory" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.498245845Z" level=warning msg="Deleting all containers under sandbox 63f26fe630a86618ccb6926337ae0f77635a06eec51dd1cb5b91c86eaefd519e since it could not be restored" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.511868714Z" level=warning msg="Could not restore sandbox 0a5caa2be3488b9c8c36d43ddc5d5a1784a28176b83143a0b5296da42812377f: failed to Statfs \"/var/run/netns/21c06552-1790-4829-956f-9ac9d4e5ebe7\": no such file or directory" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.520682871Z" level=warning msg="Deleting all containers under sandbox 0a5caa2be3488b9c8c36d43ddc5d5a1784a28176b83143a0b5296da42812377f since it could not be restored" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.539444908Z" level=warning msg="Could not restore sandbox 1ffc39284205644ee2d20c66b8df819acc6bda29c81ecc932ae2237e02a208e7: failed to Statfs \"/var/run/netns/6fa2a531-712f-4786-92e8-d0a01014b363\": no such file or directory" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.547253451Z" level=warning msg="Deleting all containers under sandbox 1ffc39284205644ee2d20c66b8df819acc6bda29c81ecc932ae2237e02a208e7 since it could not be restored" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.565457865Z" level=warning msg="Could not restore sandbox f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3: failed to Statfs \"/var/run/netns/b93696a6-c824-4b48-ae0e-2e8a183d9b47\": no such file or directory" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.571437858Z" level=warning msg="Deleting all containers under sandbox f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3 since it could not be restored" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.606830226Z" level=warning msg="Could not restore sandbox d768276def4b37ab70feeee6455eb37f4fe6716bad6df283af0049c5724973e7: failed to Statfs \"/var/run/netns/0f6c1c50-ff33-4781-9e31-36c023f9aee3\": no such file or directory" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.615870477Z" level=warning msg="Deleting all containers under sandbox d768276def4b37ab70feeee6455eb37f4fe6716bad6df283af0049c5724973e7 since it could not be restored" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.623448840Z" level=warning msg="Could not restore sandbox 8e2d55780ef7738a05fee323e5c71bd87390c87bd4b55a419d3d7362eb940075: failed to Statfs \"/var/run/netns/74a8e5e6-0f86-443d-a641-ea34c5ca1020\": no such file or directory" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.629182533Z" level=warning msg="Deleting all containers under sandbox 8e2d55780ef7738a05fee323e5c71bd87390c87bd4b55a419d3d7362eb940075 since it could not be restored" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.640295707Z" level=warning msg="Could not restore sandbox 00fb6557a9939b9a2bc7c25ec8cc1770b954010cd506656807f39f57849767a8: failed to Statfs \"/var/run/netns/4e35a0fe-47dd-4c20-9ad8-c4b066c02651\": no such file or directory" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.646014397Z" level=warning msg="Deleting all containers under sandbox 00fb6557a9939b9a2bc7c25ec8cc1770b954010cd506656807f39f57849767a8 since it could not be restored" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.652596145Z" level=warning msg="Could not restore sandbox 080616c6ee5360407adb7a3e03767428f4727a0ee1746976a16081b2970f8d20: failed to Statfs \"/var/run/netns/7a46ccd6-048b-4d90-b81f-67262d0e7bcb\": no such file or directory" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.659071741Z" level=warning msg="Deleting all containers under sandbox 080616c6ee5360407adb7a3e03767428f4727a0ee1746976a16081b2970f8d20 since it could not be restored" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.665685194Z" level=warning msg="Could not restore sandbox 7037be6a4b21795a75245d556479196ac813418de71922918e5a443959aa8e0e: failed to Statfs \"/var/run/netns/a60a285a-5ad2-491b-88d7-9cc69d283166\": no such file or directory" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.671836885Z" level=warning msg="Deleting all containers under sandbox 7037be6a4b21795a75245d556479196ac813418de71922918e5a443959aa8e0e since it could not be restored" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.678095419Z" level=warning msg="Could not restore sandbox 33e5ecc33cac76d9879725e41d559c60e7711f7e5422fe607edbdadb5460eb11: failed to Statfs \"/var/run/netns/6bbe39cf-79d5-4f40-b2bc-384feaa91c56\": no such file or directory" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.683991922Z" level=warning msg="Deleting all containers under sandbox 33e5ecc33cac76d9879725e41d559c60e7711f7e5422fe607edbdadb5460eb11 since it could not be restored" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.724100686Z" level=warning msg="Could not restore sandbox 3f3423d5651958cff05297583dd5450abb9aa783ed6eab062e62ecac5d0396ad: failed to Statfs \"/var/run/netns/ad54c509-8e21-4800-9eb5-8c6ffab0064f\": no such file or directory" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.730181730Z" level=warning msg="Deleting all containers under sandbox 3f3423d5651958cff05297583dd5450abb9aa783ed6eab062e62ecac5d0396ad since it could not be restored" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.730974834Z" level=warning msg="Could not restore sandbox 821f01442b5e38ec5d5958577a6f01e03591ea0d8fc60b05f51cb69558bd0cb1: failed to Statfs \"/var/run/netns/79c7ca97-6054-44c7-b92a-092e2473ac81\": no such file or directory" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.737096802Z" level=warning msg="Deleting all containers under sandbox 821f01442b5e38ec5d5958577a6f01e03591ea0d8fc60b05f51cb69558bd0cb1 since it could not be restored" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.756930867Z" level=warning msg="Could not restore sandbox 8cf844a5c2263b21afe261a444f7b662d35e088b48ebd92b2f02703a28a05ec2: failed to Statfs \"/var/run/netns/879d6a09-cb72-4a96-8bb2-46ff50ba4d92\": no such file or directory" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.761551809Z" level=warning msg="Deleting all containers under sandbox 8cf844a5c2263b21afe261a444f7b662d35e088b48ebd92b2f02703a28a05ec2 since it could not be restored" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.768880134Z" level=warning msg="Could not restore sandbox bb417fcfe195906c7b26f0d9c20687f555b79c545c37fe3e4cebd9d60e23d18c: failed to Statfs \"/var/run/netns/4475bb08-045b-4d37-bc48-4d8025ed2f99\": no such file or directory" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.773957862Z" level=warning msg="Deleting all containers under sandbox bb417fcfe195906c7b26f0d9c20687f555b79c545c37fe3e4cebd9d60e23d18c since it could not be restored" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.788471772Z" level=warning msg="Could not restore sandbox a87b4ab12cf54ed2e65342c3d9c4ea51c24387aaa25fa0e287ee0befc098a632: failed to Statfs \"/var/run/netns/4c6454d7-1ab1-469f-b302-44d49f84f290\": no such file or directory" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.794137922Z" level=warning msg="Deleting all containers under sandbox a87b4ab12cf54ed2e65342c3d9c4ea51c24387aaa25fa0e287ee0befc098a632 since it could not be restored" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.811218288Z" level=warning msg="Could not restore sandbox 16c0a5a6c5f7e9eecc5461c963ba5b26d7e13b2c8571603f6e05ef86ce3505e3: failed to Statfs \"/var/run/netns/185ab30d-a13f-46e2-a94f-58153e8799e2\": no such file or directory" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.816244698Z" level=warning msg="Deleting all containers under sandbox 16c0a5a6c5f7e9eecc5461c963ba5b26d7e13b2c8571603f6e05ef86ce3505e3 since it could not be restored" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.821323467Z" level=warning msg="Could not restore sandbox 01ea8cd190538ff3846bfa2fc9b101a796149e8b3c9ebd27850d03c9ae1f894d: failed to Statfs \"/var/run/netns/b878c303-c716-4fbb-9d4d-e5c5b8d6a715\": no such file or directory" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.826229563Z" level=warning msg="Deleting all containers under sandbox 01ea8cd190538ff3846bfa2fc9b101a796149e8b3c9ebd27850d03c9ae1f894d since it could not be restored" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.868332505Z" level=warning msg="Error encountered when checking whether cri-o should wipe containers: open /var/run/crio/version: no such file or directory" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.868440900Z" level=info msg="cleanup sandbox network" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.873942843Z" level=info msg="Successfully cleaned up network for pod 7037be6a4b21795a75245d556479196ac813418de71922918e5a443959aa8e0e" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.873963193Z" level=info msg="cleanup sandbox network" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.873972158Z" level=info msg="Successfully cleaned up network for pod 080616c6ee5360407adb7a3e03767428f4727a0ee1746976a16081b2970f8d20" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.873982483Z" level=info msg="cleanup sandbox network" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.878464829Z" level=info msg="Serving metrics on :9537 via HTTP" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.878797042Z" level=info msg="Got pod network &{Name:thanos-querier-6f4f5fb7cf-rdm9l Namespace:openshift-monitoring ID:f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3 UID:7c63c9be-d1ad-44b6-8b33-09af3ee314af NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:27.878959443Z" level=info msg="Deleting pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l from CNI network \"multus-cni-network\" (type=multus)" Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Started Container Runtime Interface for OCI (CRI-O). Jan 05 17:24:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Starting Kubernetes Kubelet... Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: Flag --container-runtime has been deprecated, will be removed in 1.27 as the only valid value is 'remote' Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: Flag --cloud-provider has been deprecated, will be removed in 1.25 or later, in favor of removing cloud provider code from Kubelet. Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: Flag --provider-id has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.27. Image garbage collector will get sandbox image information from CRI. Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.248433 2927 server.go:200] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: Flag --container-runtime has been deprecated, will be removed in 1.27 as the only valid value is 'remote' Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: Flag --cloud-provider has been deprecated, will be removed in 1.25 or later, in favor of removing cloud provider code from Kubelet. Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: Flag --provider-id has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.27. Image garbage collector will get sandbox image information from CRI. Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250401 2927 flags.go:64] FLAG: --add-dir-header="false" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250419 2927 flags.go:64] FLAG: --address="0.0.0.0" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250427 2927 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250435 2927 flags.go:64] FLAG: --alsologtostderr="false" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250440 2927 flags.go:64] FLAG: --anonymous-auth="true" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250447 2927 flags.go:64] FLAG: --application-metrics-count-limit="100" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250453 2927 flags.go:64] FLAG: --authentication-token-webhook="false" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250458 2927 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250465 2927 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250472 2927 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250477 2927 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250482 2927 flags.go:64] FLAG: --azure-container-registry-config="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250487 2927 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250493 2927 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250499 2927 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250505 2927 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250511 2927 flags.go:64] FLAG: --cgroup-root="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250516 2927 flags.go:64] FLAG: --cgroups-per-qos="true" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250521 2927 flags.go:64] FLAG: --client-ca-file="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250526 2927 flags.go:64] FLAG: --cloud-config="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250531 2927 flags.go:64] FLAG: --cloud-provider="external" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250536 2927 flags.go:64] FLAG: --cluster-dns="[]" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250543 2927 flags.go:64] FLAG: --cluster-domain="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250548 2927 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250554 2927 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250560 2927 flags.go:64] FLAG: --container-log-max-files="5" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250566 2927 flags.go:64] FLAG: --container-log-max-size="10Mi" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250572 2927 flags.go:64] FLAG: --container-runtime="remote" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250577 2927 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250582 2927 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250589 2927 flags.go:64] FLAG: --containerd-namespace="k8s.io" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250595 2927 flags.go:64] FLAG: --contention-profiling="false" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250600 2927 flags.go:64] FLAG: --cpu-cfs-quota="true" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250605 2927 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250611 2927 flags.go:64] FLAG: --cpu-manager-policy="none" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250616 2927 flags.go:64] FLAG: --cpu-manager-policy-options="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.250624 2927 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251226 2927 flags.go:64] FLAG: --enable-controller-attach-detach="true" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251242 2927 flags.go:64] FLAG: --enable-debugging-handlers="true" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251248 2927 flags.go:64] FLAG: --enable-load-reader="false" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251254 2927 flags.go:64] FLAG: --enable-server="true" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251260 2927 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251267 2927 flags.go:64] FLAG: --event-burst="10" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251273 2927 flags.go:64] FLAG: --event-qps="5" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251278 2927 flags.go:64] FLAG: --event-storage-age-limit="default=0" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251291 2927 flags.go:64] FLAG: --event-storage-event-limit="default=0" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251297 2927 flags.go:64] FLAG: --eviction-hard="imagefs.available<15%,memory.available<100Mi,nodefs.available<10%,nodefs.inodesFree<5%" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251314 2927 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251319 2927 flags.go:64] FLAG: --eviction-minimum-reclaim="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251325 2927 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251331 2927 flags.go:64] FLAG: --eviction-soft="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251336 2927 flags.go:64] FLAG: --eviction-soft-grace-period="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251341 2927 flags.go:64] FLAG: --exit-on-lock-contention="false" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251347 2927 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251352 2927 flags.go:64] FLAG: --experimental-mounter-path="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251357 2927 flags.go:64] FLAG: --fail-swap-on="true" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251363 2927 flags.go:64] FLAG: --feature-gates="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251370 2927 flags.go:64] FLAG: --file-check-frequency="20s" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251375 2927 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251380 2927 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251387 2927 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251393 2927 flags.go:64] FLAG: --healthz-port="10248" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251399 2927 flags.go:64] FLAG: --help="false" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251405 2927 flags.go:64] FLAG: --hostname-override="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251411 2927 flags.go:64] FLAG: --housekeeping-interval="10s" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251416 2927 flags.go:64] FLAG: --http-check-frequency="20s" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251422 2927 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251427 2927 flags.go:64] FLAG: --image-credential-provider-config="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251433 2927 flags.go:64] FLAG: --image-gc-high-threshold="85" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251438 2927 flags.go:64] FLAG: --image-gc-low-threshold="80" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251443 2927 flags.go:64] FLAG: --image-service-endpoint="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251447 2927 flags.go:64] FLAG: --iptables-drop-bit="15" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251453 2927 flags.go:64] FLAG: --iptables-masquerade-bit="14" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251458 2927 flags.go:64] FLAG: --keep-terminated-pod-volumes="false" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251464 2927 flags.go:64] FLAG: --kernel-memcg-notification="false" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251470 2927 flags.go:64] FLAG: --kube-api-burst="10" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251475 2927 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251481 2927 flags.go:64] FLAG: --kube-api-qps="5" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251486 2927 flags.go:64] FLAG: --kube-reserved="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251492 2927 flags.go:64] FLAG: --kube-reserved-cgroup="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251497 2927 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251503 2927 flags.go:64] FLAG: --kubelet-cgroups="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251508 2927 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251513 2927 flags.go:64] FLAG: --lock-file="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251519 2927 flags.go:64] FLAG: --log-backtrace-at=":0" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251525 2927 flags.go:64] FLAG: --log-cadvisor-usage="false" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251530 2927 flags.go:64] FLAG: --log-dir="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251536 2927 flags.go:64] FLAG: --log-file="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251541 2927 flags.go:64] FLAG: --log-file-max-size="1800" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251547 2927 flags.go:64] FLAG: --log-flush-frequency="5s" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251552 2927 flags.go:64] FLAG: --log-json-info-buffer-size="0" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251561 2927 flags.go:64] FLAG: --log-json-split-stream="false" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251567 2927 flags.go:64] FLAG: --logging-format="text" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251574 2927 flags.go:64] FLAG: --logtostderr="true" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251579 2927 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251586 2927 flags.go:64] FLAG: --make-iptables-util-chains="true" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251592 2927 flags.go:64] FLAG: --manifest-url="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251597 2927 flags.go:64] FLAG: --manifest-url-header="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251604 2927 flags.go:64] FLAG: --master-service-namespace="default" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251610 2927 flags.go:64] FLAG: --max-housekeeping-interval="15s" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251616 2927 flags.go:64] FLAG: --max-open-files="1000000" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251622 2927 flags.go:64] FLAG: --max-pods="110" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251628 2927 flags.go:64] FLAG: --maximum-dead-containers="-1" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251656 2927 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251661 2927 flags.go:64] FLAG: --memory-manager-policy="None" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251666 2927 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251672 2927 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251678 2927 flags.go:64] FLAG: --node-ip="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251683 2927 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/worker=,node.openshift.io/os_id=rhcos" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251695 2927 flags.go:64] FLAG: --node-status-max-images="50" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251700 2927 flags.go:64] FLAG: --node-status-update-frequency="10s" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251706 2927 flags.go:64] FLAG: --one-output="false" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251713 2927 flags.go:64] FLAG: --oom-score-adj="-999" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251719 2927 flags.go:64] FLAG: --pod-cidr="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251724 2927 flags.go:64] FLAG: --pod-infra-container-image="registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251733 2927 flags.go:64] FLAG: --pod-manifest-path="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251738 2927 flags.go:64] FLAG: --pod-max-pids="-1" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251754 2927 flags.go:64] FLAG: --pods-per-core="0" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251760 2927 flags.go:64] FLAG: --port="10250" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251765 2927 flags.go:64] FLAG: --protect-kernel-defaults="false" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251770 2927 flags.go:64] FLAG: --provider-id="alicloud://us-east-1.i-0xif5dui7khr4rm1wp68" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251776 2927 flags.go:64] FLAG: --qos-reserved="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251782 2927 flags.go:64] FLAG: --read-only-port="10255" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251787 2927 flags.go:64] FLAG: --register-node="true" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251793 2927 flags.go:64] FLAG: --register-schedulable="true" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251798 2927 flags.go:64] FLAG: --register-with-taints="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251807 2927 flags.go:64] FLAG: --registry-burst="10" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251814 2927 flags.go:64] FLAG: --registry-qps="5" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251819 2927 flags.go:64] FLAG: --reserved-cpus="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251825 2927 flags.go:64] FLAG: --reserved-memory="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251831 2927 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251837 2927 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251842 2927 flags.go:64] FLAG: --rotate-certificates="false" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251848 2927 flags.go:64] FLAG: --rotate-server-certificates="false" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251853 2927 flags.go:64] FLAG: --runonce="false" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251858 2927 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251864 2927 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251869 2927 flags.go:64] FLAG: --seccomp-default="false" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251874 2927 flags.go:64] FLAG: --serialize-image-pulls="true" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251880 2927 flags.go:64] FLAG: --skip-headers="false" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251885 2927 flags.go:64] FLAG: --skip-log-headers="false" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251890 2927 flags.go:64] FLAG: --stderrthreshold="2" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251897 2927 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251912 2927 flags.go:64] FLAG: --storage-driver-db="cadvisor" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251918 2927 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251923 2927 flags.go:64] FLAG: --storage-driver-password="root" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251929 2927 flags.go:64] FLAG: --storage-driver-secure="false" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251934 2927 flags.go:64] FLAG: --storage-driver-table="stats" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251940 2927 flags.go:64] FLAG: --storage-driver-user="root" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251945 2927 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251950 2927 flags.go:64] FLAG: --sync-frequency="1m0s" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.251955 2927 flags.go:64] FLAG: --system-cgroups="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.255051 2927 flags.go:64] FLAG: --system-reserved="cpu=500m,ephemeral-storage=1Gi,memory=1Gi" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.255165 2927 flags.go:64] FLAG: --system-reserved-cgroup="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.255214 2927 flags.go:64] FLAG: --tls-cert-file="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.255259 2927 flags.go:64] FLAG: --tls-cipher-suites="[]" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.255318 2927 flags.go:64] FLAG: --tls-min-version="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.255367 2927 flags.go:64] FLAG: --tls-private-key-file="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.255421 2927 flags.go:64] FLAG: --topology-manager-policy="none" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.255466 2927 flags.go:64] FLAG: --topology-manager-scope="container" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.255508 2927 flags.go:64] FLAG: --v="2" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.255554 2927 flags.go:64] FLAG: --version="false" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.255599 2927 flags.go:64] FLAG: --vmodule="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.255671 2927 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.255717 2927 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.255948 2927 feature_gate.go:246] feature gates: &{map[APIPriorityAndFairness:true CSIMigrationAzureFile:false CSIMigrationvSphere:false DownwardAPIHugePages:true RotateKubeletServerCertificate:true]} Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.260928 2927 server.go:413] "Kubelet version" kubeletVersion="v1.25.4+77bec7a" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.260945 2927 server.go:415] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.260999 2927 feature_gate.go:246] feature gates: &{map[APIPriorityAndFairness:true CSIMigrationAzureFile:false CSIMigrationvSphere:false DownwardAPIHugePages:true RotateKubeletServerCertificate:true]} Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.261087 2927 feature_gate.go:246] feature gates: &{map[APIPriorityAndFairness:true CSIMigrationAzureFile:false CSIMigrationvSphere:false DownwardAPIHugePages:true RotateKubeletServerCertificate:true]} Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.261197 2927 server.go:825] "Client rotation is on, will bootstrap in background" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.265307 2927 bootstrap.go:84] "Current kubeconfig file contents are still valid, no bootstrap necessary" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.265400 2927 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.265716 2927 server.go:882] "Starting client certificate rotation" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.265733 2927 certificate_manager.go:270] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.267250 2927 certificate_manager.go:270] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2023-01-06 06:46:04 +0000 UTC, rotation deadline is 2023-01-06 01:45:54.721245112 +0000 UTC Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.267269 2927 certificate_manager.go:270] kubernetes.io/kube-apiserver-client-kubelet: Waiting 8h21m26.453978404s for next certificate rotation Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.275352 2927 dynamic_cafile_content.go:119] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.275862 2927 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.276753 2927 manager.go:163] cAdvisor running in container: "/system.slice/kubelet.service" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.282060 2927 fs.go:133] Filesystem UUIDs: map[249B-6B8C:/dev/vda2 7e283ec8-4fd4-48d9-bedf-150e5fbbb374:/dev/vda3 fda260a9-581c-4391-b124-5cfcba02ef65:/dev/vda4] Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.282077 2927 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /sys/fs/cgroup:{mountpoint:/sys/fs/cgroup major:0 minor:25 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:43 fsType:tmpfs blockSize:0}] Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.282113 2927 nvidia.go:54] NVIDIA GPU metrics disabled Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.283870 2927 manager.go:212] Machine: {Timestamp:2023-01-05 17:24:28.28370494 +0000 UTC m=+0.308546029 CPUVendorID:GenuineIntel NumCores:2 NumPhysicalCores:1 NumSockets:1 CpuFrequency:2500000 MemoryCapacity:8192073728 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:2bf656612c3249b68a06a3657cb09103 SystemUUID:2bf65661-2c32-49b6-8a06-a3657cb09103 BootID:75e38bff-23b1-438d-a58e-dbb6f9d5ea7a Filesystems:[{Device:/sys/fs/cgroup DeviceMajor:0 DeviceMinor:25 Capacity:4096036864 Type:vfs Inodes:1000009 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:128300593152 Type:vfs Inodes:62651840 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:43 Capacity:4096036864 Type:vfs Inodes:1000009 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:4096036864 Type:vfs Inodes:1000009 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:4096036864 Type:vfs Inodes:1000009 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:128849018880 Scheduler:mq-deadline}] NetworkDevices:[{Name:br-ex MacAddress:00:16:3e:00:58:a6 Speed:0 Mtu:1500} {Name:ens5 MacAddress:00:16:3e:00:58:a6 Speed:-1 Mtu:1500} {Name:ovs-system MacAddress:ce:46:b4:ae:64:5d Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:8192073728 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0 1] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0}] Caches:[{Id:0 Size:37486592 Type:Unified Level:3}]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.283996 2927 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.284192 2927 manager.go:228] Version: {KernelVersion:4.18.0-372.39.1.el8_6.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 412.86.202212170457-0 (Ootpa) DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.284677 2927 container_manager_linux.go:262] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.284774 2927 container_manager_linux.go:267] "Creating Container Manager object based on Node Config" nodeConfig={RuntimeCgroupsName:/system.slice/crio.service SystemCgroupsName:/system.slice KubeletCgroupsName: KubeletOOMScoreAdj:-999 ContainerRuntime: CgroupsPerQOS:true CgroupRoot:/ CgroupDriver:systemd KubeletRootDir:/var/lib/kubelet ProtectKernelDefaults:false NodeAllocatableConfig:{KubeReservedCgroupName: SystemReservedCgroupName: ReservedSystemCPUs: EnforceNodeAllocatable:map[pods:{}] KubeReserved:map[] SystemReserved:map[cpu:{i:{value:500 scale:-3} d:{Dec:} s:500m Format:DecimalSI} ephemeral-storage:{i:{value:1073741824 scale:0} d:{Dec:} s:1Gi Format:BinarySI} memory:{i:{value:1073741824 scale:0} d:{Dec:} s:1Gi Format:BinarySI}] HardEvictionThresholds:[{Signal:nodefs.inodesFree Operator:LessThan Value:{Quantity: Percentage:0.05} GracePeriod:0s MinReclaim:} {Signal:imagefs.available Operator:LessThan Value:{Quantity: Percentage:0.15} GracePeriod:0s MinReclaim:} {Signal:memory.available Operator:LessThan Value:{Quantity:100Mi Percentage:0} GracePeriod:0s MinReclaim:} {Signal:nodefs.available Operator:LessThan Value:{Quantity: Percentage:0.1} GracePeriod:0s MinReclaim:}]} QOSReserved:map[] ExperimentalCPUManagerPolicy:none ExperimentalCPUManagerPolicyOptions:map[] ExperimentalTopologyManagerScope:container ExperimentalCPUManagerReconcilePeriod:10s ExperimentalMemoryManagerPolicy:None ExperimentalMemoryManagerReservedMemory:[] ExperimentalPodPidsLimit:4096 EnforceCPULimits:true CPUCFSQuotaPeriod:100ms ExperimentalTopologyManagerPolicy:none} Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.284800 2927 topology_manager.go:134] "Creating topology manager with policy per scope" topologyPolicyName="none" topologyScopeName="container" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.284817 2927 container_manager_linux.go:302] "Creating device plugin manager" devicePluginEnabled=true Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.284831 2927 manager.go:127] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.284845 2927 server.go:64] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.284985 2927 state_mem.go:36] "Initialized new in-memory state store" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.285043 2927 util_unix.go:104] "Using this endpoint is deprecated, please consider using full URL format" endpoint="/var/run/crio/crio.sock" URL="unix:///var/run/crio/crio.sock" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.294521 2927 remote_runtime.go:139] "Using CRI v1 runtime API" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.294544 2927 util_unix.go:104] "Using this endpoint is deprecated, please consider using full URL format" endpoint="/var/run/crio/crio.sock" URL="unix:///var/run/crio/crio.sock" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.315275 2927 remote_image.go:95] "Using CRI v1 image API" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.315313 2927 server.go:1136] "Using root directory" path="/var/lib/kubelet" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.317301 2927 kubelet.go:393] "Attempting to sync node with API server" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.317326 2927 kubelet.go:282] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.317356 2927 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.317370 2927 kubelet.go:293] "Adding apiserver pod source" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.317470 2927 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.319317 2927 kuberuntime_manager.go:240] "Container runtime initialized" containerRuntime="cri-o" version="1.25.1-5.rhaos4.12.git6005903.el8" apiVersion="v1" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 17:24:28.319981 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 17:24:28.320114 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.320137 2927 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 17:24:28.320495 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 17:24:28.320569 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.322540 2927 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.322569 2927 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/rbd" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.322582 2927 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/aws-ebs" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.322593 2927 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/gce-pd" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.322604 2927 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/cinder" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.322615 2927 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/azure-disk" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.322628 2927 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/azure-file" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.322659 2927 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/vsphere-volume" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.323017 2927 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.323030 2927 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.323041 2927 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.323053 2927 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.323065 2927 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/secret" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.323082 2927 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.323093 2927 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/glusterfs" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.323104 2927 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/cephfs" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.323116 2927 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.323131 2927 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/fc" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.323142 2927 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.323154 2927 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/projected" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.323164 2927 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.323775 2927 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/csi" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.323944 2927 server.go:1175] "Started kubelet" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Started Kubernetes Kubelet. Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.325235 2927 server.go:155] "Starting to listen" address="0.0.0.0" port=10250 Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.326873 2927 server.go:438] "Adding debug handlers to kubelet server" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 17:24:28.328823 2927 kubelet.go:1333] "Image garbage collection failed once. Stats initialization may not have completed yet" err="failed to get imageFs info: unable to find data in memory cache" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.331870 2927 certificate_manager.go:270] kubernetes.io/kubelet-serving: Certificate rotation is enabled Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.331976 2927 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.334605 2927 certificate_manager.go:270] kubernetes.io/kubelet-serving: Certificate expiration is 2023-01-06 06:46:04 +0000 UTC, rotation deadline is 2023-01-06 03:19:21.929377205 +0000 UTC Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.334764 2927 certificate_manager.go:270] kubernetes.io/kubelet-serving: Waiting 9h54m53.594636365s for next certificate rotation Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.337361 2927 volume_manager.go:291] "The desired_state_of_world populator starts" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.337453 2927 volume_manager.go:293] "Starting Kubelet Volume Manager" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.338442 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:28.340275231Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=a885c2e3-2535-4768-969a-3567aabac631 name=/runtime.v1.ImageService/ImageStatus Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.339167 2927 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 17:24:28.345057 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 17:24:28.345182 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.349441 2927 factory.go:153] Registering CRI-O factory Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.349458 2927 factory.go:55] Registering systemd factory Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.349543 2927 factory.go:103] Registering Raw factory Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.349646 2927 manager.go:1201] Started watching for new ooms in manager Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.350446 2927 manager.go:302] Starting recovery of all containers Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 17:24:28.353485833Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=a885c2e3-2535-4768-969a-3567aabac631 name=/runtime.v1.ImageService/ImageStatus Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 17:24:28.369049 2927 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.17377995342d9680", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 323919488, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 323919488, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 17:24:28.377721 2927 controller.go:144] failed to ensure lease exists, will retry in 200ms, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.439876 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 17:24:28.443298 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.494556 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.494761 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.494863 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.494960 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 17:24:28.498951 2927 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.516268 2927 manager.go:307] Recovery completed Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 17:24:28.544062 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.598362 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 17:24:28.599845 2927 controller.go:144] failed to ensure lease exists, will retry in 400ms, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.600871 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.600899 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.600914 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.601411 2927 cpu_manager.go:214] "Starting CPU manager" policy="none" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.601430 2927 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.601447 2927 state_mem.go:36] "Initialized new in-memory state store" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.603383 2927 policy_none.go:49] "None policy: Start" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.604016 2927 memory_manager.go:168] "Starting memorymanager" policy="None" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.604064 2927 state_mem.go:35] "Initializing new in-memory state store" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.605538 2927 container_manager_linux.go:427] "Updating kernel flag" flag="vm/overcommit_memory" expectedValue=1 actualValue=0 Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.605666 2927 container_manager_linux.go:427] "Updating kernel flag" flag="kernel/panic" expectedValue=10 actualValue=0 Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.606427 2927 kubelet_network_linux.go:63] "Initialized iptables rules." protocol=IPv4 Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 17:24:28.644694 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.658083 2927 manager.go:273] "Starting Device Plugin manager" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.658132 2927 manager.go:447] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.658143 2927 server.go:77] "Starting device plugin registration server" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.658503 2927 plugin_watcher.go:52] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.658584 2927 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.658594 2927 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 17:24:28.659314 2927 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.667368 2927 kubelet_network_linux.go:63] "Initialized iptables rules." protocol=IPv6 Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.667470 2927 status_manager.go:161] "Starting to sync pod status with apiserver" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.667550 2927 kubelet.go:2033] "Starting kubelet main sync loop" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 17:24:28.667666 2927 kubelet.go:2057] "Skipping pod synchronization" err="PLEG is not healthy: pleg has yet to be successful" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 17:24:28.669946 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 17:24:28.670103 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.699491 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.700394 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.700426 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.700441 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.700462 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 17:24:28.701222 2927 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 17:24:28.745662 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 17:24:28.768767 2927 kubelet.go:2119] "SyncLoop ADD" source="file" pods=[] Jan 05 17:24:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 17:24:28.846259 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:29.215839 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:29.270581 2927 controller.go:144] failed to ensure lease exists, will retry in 800ms, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:29.316893 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:29.371019 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:24:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:29.372060 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:24:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:29.372089 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:24:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:29.372103 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:24:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:29.372127 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:24:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:29.372773 2927 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:24:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:24:29.408508 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:29.408556 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:29.417757 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:24:29.474571 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:29.474616 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:29.517929 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:29.614157 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:29.618294 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:29.718731 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:24:29.787667 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:29.787700 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:29.819053 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:29.919481 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:24:29.966194 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:29.966240 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:30.020542 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:30.071470 2927 controller.go:144] failed to ensure lease exists, will retry in 1.6s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:30.120909 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:30.172984 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:24:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:30.174192 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:24:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:30.174219 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:24:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:30.174233 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:24:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:30.174256 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:24:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:30.174976 2927 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:24:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:30.221474 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:30.321981 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:30.422494 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:30.523022 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:30.614405 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:30.623594 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:30.724104 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:30.824625 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:30.925136 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:31.025510 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:31.126108 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:31.226641 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:31.327167 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:31.427696 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:31.528226 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:31.614649 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:31.628853 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:31.672584 2927 controller.go:144] failed to ensure lease exists, will retry in 3.2s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:31.715481 2927 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.17377995342d9680", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 323919488, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 323919488, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:24:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:31.729693 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:24:31.749518 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:31.749550 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:31.775592 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:24:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:31.776626 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:24:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:31.776651 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:24:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:31.776664 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:24:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:31.776690 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:24:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:31.777404 2927 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:24:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:31.829860 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:31.930388 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:32.030878 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:32.131517 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:24:32.164439 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:32.164473 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:32.231863 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:32.332395 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:32.432925 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:24:32.491883 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:32.491935 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:32.533293 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:32.614586 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:24:32.623268 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:32.623298 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:32.633461 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:32.733981 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:32.834506 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:32.935050 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:33.035406 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:33.135939 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:33.236465 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:33.336985 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:33.437513 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:33.538185 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:33.614567 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:33.638560 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:33.739055 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:33.839516 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:33.939764 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:34.039938 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:34.140522 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:34.241045 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:34.341512 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:34.442033 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:34.542497 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:34.613855 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:34.643059 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:34.743522 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:34.844044 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:34.873850 2927 controller.go:144] failed to ensure lease exists, will retry in 6.4s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:34.944101 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:34.978307 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:24:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:34.979327 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:24:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:34.979354 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:24:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:34.979368 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:24:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:34.979390 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:24:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:34.980096 2927 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:24:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:35.044500 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:35.144981 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:35.245489 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:35.345985 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:35.446483 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:35.547036 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:35.614175 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:35.647364 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:35.747849 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:35.848326 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:35.948642 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:36.049200 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:36.149687 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:36.250186 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:36.350666 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:24:36.389658 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:36.389692 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:36.451216 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:36.551829 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:36.613966 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:36.652168 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:36.752640 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:36.853115 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:36.953583 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:37.054187 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:37.154663 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:37.255114 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:37.355551 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:37.456010 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:37.556611 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:37.614676 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:37.656875 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:37.757367 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:24:37.786607 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:37.786648 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:37.858027 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:37.958540 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:24:37.973352 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:37.973383 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:38.058859 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:38.159270 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:38.259672 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:38.360128 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:24:38.371012 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:38.371045 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:38.460445 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:38.561390 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:38.614574 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:38.661772 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:38.762245 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:38.862690 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:38.928697 2927 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:38.963510 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:39.064051 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:39.164505 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:39.264935 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:39.365410 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:39.465938 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:39.566131 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:39.614278 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:39.666481 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:39.766794 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:39.867263 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:39.967609 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:40.068273 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:40.168697 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:40.269181 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:40.369644 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:40.470129 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:40.570550 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:40.614840 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:40.671045 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:40.771480 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:40.871931 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:40.972241 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:41.072841 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:41.173111 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:41.273440 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:41.275279 2927 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:41.373680 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:41.380780 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:24:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:41.381926 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:24:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:41.381961 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:24:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:41.381976 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:24:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:41.382000 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:24:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:41.382757 2927 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:24:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:41.474227 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:41.574998 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:41.614183 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:41.675192 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:41.717274 2927 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.17377995342d9680", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 323919488, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 323919488, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:24:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:41.775498 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:41.875845 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:41.976728 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:42.077138 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:42.177526 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:42.277987 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:42.378934 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:42.479442 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:42.580192 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:42.613936 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:42.680664 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:42.780728 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:42.881131 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:42.981313 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:43.081927 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:43.182376 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:43.282859 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:43.383898 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:43.484207 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:43.584928 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:43.613758 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:43.685932 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:43.786255 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:43.886692 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:43.987503 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:44.088103 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:44.189176 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:44.289249 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:44.389594 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:44.490065 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:44.590610 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:44.614615 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:44.690845 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:44.791169 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:44.891644 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:44.992064 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:45.092598 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:45.193054 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:45.293371 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:45.393872 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:45.494392 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:24:45.509224 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:45.509273 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:45.595416 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:45.614938 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:45.695931 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:45.796436 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:45.896931 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:45.997442 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:46.098048 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:46.198506 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:46.298702 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:46.399299 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:46.499768 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:46.600433 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:46.614388 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:46.700634 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:46.801186 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:46.901420 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:47.001899 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:47.102402 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:47.202938 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:47.303475 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:24:47.339687 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:47.339745 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:24:47.363506 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:47.363554 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:24:47.381240 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:47.381282 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:47.404437 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:47.505489 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:47.606181 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:47.614222 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:47.706427 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:47.806830 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:47.907321 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:48.007485 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:48.107882 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:48.208919 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:48.276780 2927 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:48.309993 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:48.382802 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:24:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:48.384194 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:24:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:48.384220 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:24:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:48.384235 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:24:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:48.384257 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:24:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:48.384976 2927 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:24:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:48.410929 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:48.511380 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:48.612184 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:48.614003 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:48.712456 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:48.812872 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:48.913306 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:48.929446 2927 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:49.013487 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:49.113865 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:49.214178 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:49.314526 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:49.414931 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:49.515354 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:49.614597 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:49.615622 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:49.716048 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:49.816568 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:49.917049 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:50.017455 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:50.118027 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:50.218432 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:50.318844 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:50.419248 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:50.519663 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:50.613852 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:50.619927 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:50.720421 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:50.820930 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:50.921438 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:51.021932 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:51.122514 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:51.223028 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:51.323482 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:51.424005 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:51.524477 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:51.614667 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:51.624934 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:51.718970 2927 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.17377995342d9680", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 323919488, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 323919488, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:24:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:51.725119 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:51.825631 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:51.926140 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:52.026497 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:52.127104 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:52.227630 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:52.328110 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:52.428627 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:52.529156 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:52.614196 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:52.629501 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:52.729727 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:52.830240 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:52.930689 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:53.031139 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:53.131703 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:53.232183 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:53.332689 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:53.433058 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:53.533722 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:53.613831 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:53.634160 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:53.734655 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:53.835267 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:53.936169 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:54.036485 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:54.137046 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:54.237498 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:54.337991 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:54.438493 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:54.539001 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:54.614218 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:54.639488 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:54.739981 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:54.840488 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:54.941048 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:55.041359 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:55.141919 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:55.242416 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:55.278253 2927 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:55.342680 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:55.385797 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:24:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:55.386864 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:24:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:55.386897 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:24:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:55.386922 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:24:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:55.386945 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:24:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:55.387675 2927 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:24:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:55.443123 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:55.543631 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:55.614730 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:55.644047 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:55.744492 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:55.844995 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:55.945613 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:56.045873 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:56.146360 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:56.246843 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:56.347295 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:56.447717 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:56.548303 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:56.614394 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:56.648709 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:56.749183 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:56.849520 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:56.950127 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:57.050305 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:57.150729 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:57.251189 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:57.351693 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:57.452196 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:57.552793 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:57.613859 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:57.653184 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:57.753654 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:57.854130 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:57.954874 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:58.055195 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:58.155680 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:58.256168 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:58.356639 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:58.457105 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:58.558034 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:58.614108 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:58.658511 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:58.759055 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:58.859889 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:58.930283 2927 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:58.960668 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:59.060751 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:59.161223 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:59.261698 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:59.362193 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:59.462643 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:59.563734 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:24:59.613755 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:24:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:59.664054 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:59.764484 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:59.864835 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:24:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:24:59.965541 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:00.065935 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:00.166240 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:00.266720 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:25:00.334835 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:00.334887 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:00.366984 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:00.467199 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:00.567557 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:00.614848 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:00.668113 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:00.768451 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:00.868784 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:00.969336 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:01.069639 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:01.169994 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:01.270483 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:01.370933 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:01.471429 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:01.572057 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:01.614172 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:01.672490 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:01.720581 2927 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.17377995342d9680", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 323919488, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 323919488, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:25:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:01.772791 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:01.872933 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:01.973071 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:02.073847 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:02.174325 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:02.274814 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:02.279659 2927 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:02.375886 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:02.387990 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:25:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:02.389190 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:25:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:02.389220 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:25:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:02.389234 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:25:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:02.389258 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:25:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:02.389997 2927 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:25:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:02.476302 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:02.576863 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:02.613886 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:02.677198 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:02.777663 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:02.878139 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:02.978649 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:03.079052 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:03.179496 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:03.279985 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:03.380490 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:03.480986 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:03.581673 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:03.614649 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:03.681938 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:03.782438 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:03.882928 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:03.983808 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:04.084439 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:04.184832 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:04.285169 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:04.385639 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:04.486121 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:04.586613 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:04.614573 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:04.686891 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:04.787398 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:04.887879 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:04.988338 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:05.088699 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:05.189192 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:25:05.236325 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:05.236375 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:05.289507 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:05.389995 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:05.490322 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:05.590809 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:05.614761 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:05.691052 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:05.791508 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:05.891992 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:05.992472 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:06.092907 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:06.193380 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:06.293722 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:25:06.379852 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:06.379898 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:06.394039 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:06.494518 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:06.595119 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:06.614088 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:06.695431 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:06.795932 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:06.896442 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:06.996884 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:07.097471 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:07.197558 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:07.298130 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:07.398479 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:07.498996 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:07.599615 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:07.614625 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:07.699823 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:07.800119 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:07.900185 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:08.000839 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:08.101419 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:08.201786 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:08.302303 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:08.402612 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:08.503310 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:08.604115 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:08.614221 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:08.704566 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:08.805125 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:08.905660 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:08.931008 2927 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:25:08.966989 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:08.967042 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:09.005934 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:09.106569 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:09.207057 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:09.280983 2927 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:09.308099 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:09.390374 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:25:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:09.391547 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:25:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:09.391576 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:25:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:09.391589 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:25:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:09.391611 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:25:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:09.392313 2927 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:25:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:09.408511 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:09.508929 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:09.609492 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:09.614395 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:09.709645 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:09.810107 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:09.910395 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:10.010915 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:10.111490 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:10.211989 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:10.312372 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:10.412832 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:10.513316 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:10.613939 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:10.614673 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:10.714998 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:10.815532 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:10.915815 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:11.016246 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:11.116664 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:11.216986 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:11.317313 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:11.418354 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:11.518594 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:11.614738 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:11.618817 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:11.719010 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:11.721828 2927 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.17377995342d9680", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 323919488, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 323919488, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:25:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:11.819264 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:11.920112 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:12.020676 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:12.121301 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:12.221691 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:12.321934 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:12.422388 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:12.522589 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:12.614805 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:12.622888 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:12.723402 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:12.823870 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:12.924334 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:13.024724 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:13.125300 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:25:13.169570078Z" level=error msg="Failed to cleanup (probably retrying): failed to destroy network for pod sandbox k8s_thanos-querier-6f4f5fb7cf-rdm9l_openshift-monitoring_7c63c9be-d1ad-44b6-8b33-09af3ee314af_0(f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3): error removing pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/7c63c9be-d1ad-44b6-8b33-09af3ee314af]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" Jan 05 09:25:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:13.225650 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:13.326051 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:13.426510 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:13.527110 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:13.614270 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:13.627697 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:25:13.670209452Z" level=info msg="cleanup sandbox network" Jan 05 09:25:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:25:13.670530829Z" level=info msg="Got pod network &{Name:thanos-querier-6f4f5fb7cf-rdm9l Namespace:openshift-monitoring ID:f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3 UID:7c63c9be-d1ad-44b6-8b33-09af3ee314af NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:25:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:25:13.670707844Z" level=info msg="Deleting pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:25:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:13.727943 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:13.828311 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:13.928821 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:14.029695 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:14.130108 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:14.230614 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:14.331106 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:14.431453 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:14.532042 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:14.613946 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:14.632123 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:14.732636 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:14.832818 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:14.933316 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:15.033802 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:15.134053 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:15.234498 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:15.335003 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:15.435518 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:15.535994 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:15.614192 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:15.636434 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:15.737006 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:15.837470 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:15.937840 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:16.038176 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:16.138521 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:16.239013 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:16.281816 2927 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:16.339251 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:16.392937 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:25:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:16.394075 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:25:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:16.394104 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:25:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:16.394118 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:25:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:16.394162 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:25:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:16.394927 2927 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:25:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:16.440305 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:16.540753 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:16.613842 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:16.641202 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:16.741711 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:16.842212 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:16.942837 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:17.043156 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:17.143629 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:17.244008 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:17.344497 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:17.444995 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:17.545305 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:17.614260 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:17.645589 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:17.746055 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:17.846513 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:17.947013 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:18.047315 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:18.147807 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:18.248296 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:18.348737 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:18.449253 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:18.550065 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:18.614059 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:18.650383 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:18.750893 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:18.851381 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:18.931771 2927 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:18.951836 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:19.052188 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:19.152679 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:19.252990 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:19.353478 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:19.454518 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:19.555022 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:19.613962 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:19.655992 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:19.756493 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:19.856934 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:19.957393 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:20.057691 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:20.158168 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:20.258656 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:20.359129 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:20.459368 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:20.559939 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:20.613932 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:20.660256 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:20.760743 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:20.861224 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:20.961935 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:21.063007 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:21.163298 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:21.263749 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:21.363990 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:21.464276 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:21.564838 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:21.614757 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:21.665055 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:21.723055 2927 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.17377995342d9680", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 323919488, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 323919488, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:25:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:21.765273 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:21.865708 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:21.966437 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:22.066829 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:22.167298 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:22.267735 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:22.368056 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:22.468360 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:22.568937 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:22.614746 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:22.669991 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:22.770267 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:22.870696 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:22.971471 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:23.071831 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:23.172189 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:23.272655 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:23.283556 2927 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:23.373047 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:23.395175 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:25:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:23.396300 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:25:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:23.396329 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:25:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:23.396342 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:25:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:23.396364 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:25:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:23.397120 2927 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:25:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:23.473518 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:23.574256 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:23.614252 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:23.674574 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:23.775058 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:23.875498 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:23.976455 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:24.076685 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:24.176991 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:24.277187 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:24.377654 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:24.477931 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:24.578609 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:24.614577 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:24.678704 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:24.779185 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:24.880240 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:24.980931 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:25.081692 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:25.182175 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:25.282521 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:25.382829 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:25.483308 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:25.583706 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:25.614605 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:25.683933 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:25.784426 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:25.884908 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:25.985453 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:26.085680 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:26.186168 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:26.286632 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:26.386928 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:26.487264 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:26.587806 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:26.614730 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:26.688830 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:26.789126 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:26.889451 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:26.990183 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:27.090559 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:27.191050 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:27.291135 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:27.391623 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:27.491934 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:27.592557 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:27.614544 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:27.692878 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:27.793398 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:27.893888 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:27.994607 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:28.094784 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:28.195291 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:28.295718 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:28.395930 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:28.496549 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:28.596609 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:28.614639 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:28.696777 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:28.797283 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:28.897444 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:28.931865 2927 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:28.997931 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:29.098243 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:29.198744 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:29.299244 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:29.399737 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:29.500247 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:29.600837 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:29.614790 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:29.701114 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:29.801486 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:29.901986 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:30.002713 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:30.103051 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:30.203508 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:30.284664 2927 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:30.303853 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:30.397307 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:25:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:30.398487 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:25:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:30.398514 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:25:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:30.398528 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:25:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:30.398550 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:25:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:30.399344 2927 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:25:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:30.404421 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:30.504924 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:30.605474 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:30.613821 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:30.706014 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:30.806359 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:30.906850 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:31.007001 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:31.107839 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:31.208262 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:31.308695 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:31.409054 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:31.509524 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:31.610060 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:31.613921 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:31.710182 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:31.724071 2927 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.17377995342d9680", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 323919488, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 323919488, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:25:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:31.810378 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:31.910750 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:32.011397 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:25:32.047500 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:32.047552 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:32.111809 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:32.212122 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:32.312632 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:32.413049 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:32.513460 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:32.613929 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:32.614745 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:32.714271 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:32.814742 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:32.915199 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:33.015519 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:33.116112 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:33.216572 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:33.316909 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:33.417254 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:33.517797 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:33.614203 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:33.618382 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:33.719005 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:33.819536 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:33.920055 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:34.020379 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:34.120884 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:34.221316 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:34.321736 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:34.422105 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:34.522517 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:34.614644 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:34.622845 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:34.723380 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:34.823909 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:34.924373 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:35.024730 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:35.125182 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:35.225616 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:35.325930 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:35.426371 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:35.526781 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:35.613878 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:35.627055 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:35.727406 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:35.827861 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:35.928117 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:36.028670 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:36.129312 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:36.229813 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:36.329927 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:36.430359 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:36.530875 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:36.614007 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:36.631264 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:36.731885 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:36.832420 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:36.932932 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:37.033477 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:37.134046 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:37.234493 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:37.286343 2927 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:37.334727 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:37.399855 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:25:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:37.401049 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:25:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:37.401080 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:25:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:37.401094 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:25:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:37.401116 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:25:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:37.401822 2927 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:25:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:37.435238 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:37.535694 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:37.614544 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:37.635862 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:37.736304 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:37.836747 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:37.937113 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:38.038057 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:38.138666 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:38.239177 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:38.339510 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:38.439984 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:38.540939 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:38.614003 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:38.641322 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:25:38.721473 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:38.721522 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:38.741660 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:38.842152 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:38.932654 2927 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:38.942662 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:39.042752 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:39.143182 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:39.243675 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:39.344011 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:39.444475 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:39.544939 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:39.613892 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:39.645296 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:39.745745 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:39.846110 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:39.946497 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:40.046934 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:40.147410 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:40.247711 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:40.348179 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:40.448655 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:40.549270 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:40.614186 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:40.649609 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:40.750128 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:40.850607 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:40.951186 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:41.051659 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:41.152160 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:41.252644 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:41.353118 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:41.453603 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:41.554192 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:41.613956 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:41.655358 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:41.725445 2927 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.17377995342d9680", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 323919488, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 323919488, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:25:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:41.755657 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:41.855989 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:41.956608 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:42.057008 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:42.157297 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:42.257710 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:42.358168 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:42.458652 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:42.559135 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:42.614035 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:42.659314 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:42.759794 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:42.860256 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:42.960799 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:43.061416 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:43.161933 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:43.262418 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:43.362897 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:43.463932 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:43.564478 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:43.614396 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:43.664826 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:43.765051 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:43.865481 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:43.965555 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:44.065932 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:44.166253 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:44.266670 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:44.287494 2927 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:44.366740 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:44.402848 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:25:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:44.403924 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:25:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:44.403953 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:25:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:44.403968 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:25:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:44.403991 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:25:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:44.404684 2927 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:25:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:44.466931 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:44.567548 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:44.614467 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:44.667703 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:44.768180 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:44.868464 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:44.969048 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:45.069484 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:45.169936 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:45.270407 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:45.370740 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:45.471176 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:45.571779 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:45.613916 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:45.672359 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:25:45.727529 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:45.727583 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:45.772732 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:45.873184 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:45.973782 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:46.074240 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:46.175303 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:46.275726 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:46.375994 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:46.476478 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:46.577054 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:46.613992 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:46.677372 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:46.777843 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:46.878306 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:46.978743 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:47.079158 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:47.179489 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:47.279933 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:47.380252 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:47.480504 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:47.581109 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:47.614058 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:47.682163 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:47.782630 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:47.882920 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:47.983645 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:48.084596 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:48.185055 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:48.285480 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:48.385934 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:48.486392 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:48.587198 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:48.614233 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:48.687442 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:48.787927 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:48.888430 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:48.933688 2927 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:48.988887 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:49.089164 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:49.189625 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:49.289884 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:49.390368 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:49.490844 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:49.591423 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:49.614377 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:49.691680 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:49.792181 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:49.892656 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:49.993325 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:25:50.010261 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:50.010312 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:50.094269 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:50.194684 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:50.294932 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:50.395263 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:50.495681 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:50.596288 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:50.614252 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:50.696699 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:50.797189 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:50.897669 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:50.998022 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:51.098893 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:51.199235 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:51.288488 2927 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:51.299560 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:51.400047 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:51.405173 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:25:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:51.406261 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:25:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:51.406285 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:25:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:51.406298 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:25:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:51.406320 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:25:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:51.407036 2927 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:25:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:51.500253 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:51.600806 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:51.614764 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:51.701018 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:51.726420 2927 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.17377995342d9680", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 323919488, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 323919488, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:25:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:51.801635 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:51.901933 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:52.002530 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:52.102858 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:52.203111 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:52.304332 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:52.405101 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:52.505602 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:52.605994 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:52.613942 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:52.706362 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:52.806705 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:52.907195 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:53.007804 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:53.108819 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:53.209207 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:53.309600 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:53.409996 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:53.510447 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:53.611179 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:53.613941 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:53.711472 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:53.811928 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:53.912424 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:54.013461 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:54.113685 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:54.214056 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:54.314516 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:54.414929 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:54.515392 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:54.614591 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:54.615612 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:54.716180 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:54.816501 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:54.916988 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:55.017723 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:55.118052 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:55.218561 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:55.319049 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:55.419559 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:55.519934 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:55.614185 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:55.620250 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:55.720988 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:55.821504 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:55.922005 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:56.022794 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:56.123110 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:56.223597 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:56.324103 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:56.424482 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:56.524991 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:56.614088 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:56.625189 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:56.725765 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:56.826103 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:56.926569 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:57.027299 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:57.127559 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:57.227991 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:57.328377 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:57.428753 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:57.529160 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:57.614292 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:57.629247 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:57.729736 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:57.830745 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:57.931001 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:58.031678 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:58.132105 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:58.232609 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:58.289492 2927 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:58.332928 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:58.407867 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:25:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:58.408928 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:25:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:58.408963 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:25:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:58.408976 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:25:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:58.408995 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:25:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:58.409731 2927 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:25:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:58.433004 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:58.533674 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:58.613964 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:58.634110 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:25:58.681048628Z" level=error msg="Failed to cleanup (probably retrying): failed to destroy network for pod sandbox k8s_thanos-querier-6f4f5fb7cf-rdm9l_openshift-monitoring_7c63c9be-d1ad-44b6-8b33-09af3ee314af_0(f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3): error removing pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/7c63c9be-d1ad-44b6-8b33-09af3ee314af]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" Jan 05 09:25:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:58.734627 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:58.835323 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:58.934780 2927 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:58.935841 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:59.036250 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:59.136332 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:59.236831 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:59.337303 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:25:59.431997348Z" level=info msg="cleanup sandbox network" Jan 05 09:25:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:25:59.432257883Z" level=info msg="Got pod network &{Name:thanos-querier-6f4f5fb7cf-rdm9l Namespace:openshift-monitoring ID:f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3 UID:7c63c9be-d1ad-44b6-8b33-09af3ee314af NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:25:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:25:59.432407824Z" level=info msg="Deleting pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:25:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:59.437579 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:59.537937 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:25:59.613992 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:25:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:59.638251 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:59.738795 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:59.839299 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:25:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:25:59.939871 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:00.040150 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:00.140707 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:00.241045 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:00.341484 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:00.441816 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:00.541996 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:00.614115 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:00.642430 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:00.742935 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:00.843282 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:00.943836 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:01.043926 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:01.144484 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:01.244942 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:01.345426 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:01.445926 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:01.546434 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:01.614496 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:01.646529 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:01.727730 2927 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.17377995342d9680", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 323919488, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 323919488, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:26:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:01.746825 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:01.847301 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:01.947871 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:02.048934 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:02.149991 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:02.250097 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:02.351100 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:02.451599 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:02.552196 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:02.614136 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:02.652297 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:02.752883 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:02.853357 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:02.953930 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:03.054435 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:03.154919 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:03.255356 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:03.355827 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:03.456244 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:03.556836 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:03.614866 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:03.657188 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:03.757675 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:03.858428 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:03.959456 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:04.059915 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:04.160384 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:04.260769 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:04.361129 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:04.461613 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:04.562180 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:04.614185 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:04.662611 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:04.763052 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:04.863477 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:04.964052 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:05.064431 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:05.164889 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:05.265186 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:05.291059 2927 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:05.365480 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:05.410710 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:26:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:05.411800 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:26:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:05.411828 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:26:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:05.411841 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:26:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:05.411863 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:26:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:05.412628 2927 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:26:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:05.465774 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:05.566051 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:05.613992 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:05.666439 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:05.766927 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:05.867182 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:05.967771 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:06.067994 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:06.168426 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:06.268900 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:06.369369 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:06.469710 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:06.570423 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:06.614443 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:06.670875 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:06.771356 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:06.871827 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:06.972415 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:07.072817 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:07.173155 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:07.273616 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:07.374050 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:07.474422 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:07.575054 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:07.614006 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:07.675445 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:07.775745 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:07.876174 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:07.976763 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:08.077164 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:08.177515 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:08.277990 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:08.378300 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:08.478725 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:08.579675 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:08.614773 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:08.680023 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:08.780498 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:08.880934 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:08.935307 2927 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:08.981559 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:09.082050 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:09.182470 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:09.282931 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:09.383412 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:09.483888 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:09.584489 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:09.614432 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:09.684811 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:09.785294 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:09.885711 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:09.985798 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:10.086018 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:10.186484 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:10.286930 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:10.387407 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:10.487886 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:10.588484 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:10.614677 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:10.689117 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:10.789596 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:10.890051 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:10.990574 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:11.090847 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:11.191368 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:11.291713 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:11.392511 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:11.493063 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:11.593623 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:11.614577 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:11.694015 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:11.729022 2927 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.17377995342d9680", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 323919488, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 323919488, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:26:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:11.729095 2927 event.go:221] Unable to write event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.17377995342d9680", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 323919488, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 323919488, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}' (retry limit exceeded!) Jan 05 09:26:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:11.729811 2927 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.173779953e5bec41", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientMemory", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientMemory", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 494728257, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 494728257, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:26:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:11.795034 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:11.895471 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:11.996190 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:12.096622 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:12.197112 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:12.292364 2927 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:12.297434 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:12.397932 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:12.412680 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:26:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:12.413735 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:26:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:12.413855 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:26:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:12.413872 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:26:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:12.413897 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:26:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:12.414731 2927 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:26:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:12.498936 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:12.599541 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:12.614533 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:12.699932 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:12.800489 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:12.900933 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:13.001521 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:13.101820 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:13.202317 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:13.302811 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:13.403302 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:13.503538 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:13.604248 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:13.614213 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:26:13.631001 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:13.631049 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:13.704534 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:13.805018 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:13.905328 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:14.005481 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:14.105805 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:14.206304 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:14.306458 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:14.406898 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:14.507367 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:14.607714 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:14.614701 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:14.707930 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:14.808457 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:14.908660 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:15.009048 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:15.109244 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:15.209700 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:15.310119 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:15.410598 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:15.510847 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:15.611415 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:15.614260 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:15.711812 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:15.812310 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:15.912759 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:16.013392 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:16.113669 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:16.214060 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:16.314248 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:16.414515 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:16.514948 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:16.614214 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:16.615215 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:16.715819 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:16.816305 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:16.916798 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:17.017541 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:17.117886 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:17.218180 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:17.318679 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:17.419182 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:17.519564 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:17.614670 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:17.620535 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:17.720931 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:17.821390 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:17.921890 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:18.022558 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:18.122717 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:18.222989 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:18.323485 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:18.423833 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:18.524489 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:18.614475 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:18.625560 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:18.726283 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:18.826796 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:18.927298 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:18.935362 2927 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:19.027849 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:19.128158 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:19.228587 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:19.293473 2927 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:19.328915 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:19.415358 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:26:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:19.416537 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:26:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:19.416570 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:26:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:19.416583 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:26:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:19.416605 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:26:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:19.417348 2927 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:26:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:19.429488 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:19.529857 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:19.613998 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:19.630082 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:19.730603 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:19.831102 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:19.931562 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:20.032247 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:20.132444 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:20.232826 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:20.333062 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:20.433391 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:20.533844 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:20.613780 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:20.634864 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:20.735544 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:20.755403 2927 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.173779953e5bec41", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientMemory", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientMemory", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 494728257, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 494728257, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:26:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:20.835861 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:20.936180 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:21.036839 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:21.137109 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:21.237557 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:21.337926 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:21.438375 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:21.538774 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:21.613892 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:21.638945 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:21.739567 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:21.840599 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:21.941179 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:22.041853 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:22.142187 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:22.242674 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:22.343048 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:22.443500 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:22.543761 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:22.614809 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:22.644057 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:22.744664 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:22.844991 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:22.945245 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:23.045750 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:23.145935 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:23.246444 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:23.346848 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:23.447188 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:23.547936 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:23.613991 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:23.648315 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:23.748916 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:23.849389 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:23.950411 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:24.051182 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:24.151368 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:24.251845 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:24.352303 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:24.452616 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:24.553185 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:24.614086 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:24.653325 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:24.753928 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:24.854438 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:24.955042 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:25.055775 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:25.155938 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:25.256438 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:25.356909 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:25.457391 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:25.557989 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:25.614022 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:25.658373 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:25.758996 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:25.859472 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:25.960060 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:26.060790 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:26.161057 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:26.261491 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:26.294259 2927 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:26.361689 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:26.417789 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:26:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:26.418978 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:26:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:26.419007 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:26:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:26.419020 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:26:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:26.419042 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:26:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:26.419731 2927 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:26:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:26.461934 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:26.562336 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:26.614305 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:26.662692 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:26.763298 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:26.863715 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:26.964238 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:27.064942 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:27.165152 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:27.265513 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:27.365725 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:27.466187 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:27.566551 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:27.614227 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:27.667543 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:27.768158 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:27.868607 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:27.969270 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:28.069925 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:28.170937 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:26:28.266281 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:28.266334 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:28.271342 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:28.371691 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:28.472176 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:26:28.494184 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:28.494237 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:28.572874 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:28.614283 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:28.673487 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:28.773775 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:28.874434 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:28.936415 2927 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:28.975521 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:29.076520 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:29.176993 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:29.277329 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:29.377809 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:29.478185 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:29.578659 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:29.614577 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:29.678989 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:29.779473 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:29.879930 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:29.980526 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:30.080935 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:30.181411 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:30.281886 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:30.382353 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:30.482544 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:30.583199 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:30.613985 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:30.683425 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:30.756504 2927 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.173779953e5bec41", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientMemory", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientMemory", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 494728257, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 494728257, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:26:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:30.783594 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:30.884046 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:30.984315 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:31.084754 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:31.185294 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:31.285711 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:31.386168 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:31.486633 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:31.587237 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:31.614242 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:31.687549 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:31.787865 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:31.888344 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:31.988936 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:32.088986 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:32.189524 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:32.289700 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:32.389993 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:32.490479 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:32.591058 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:32.614067 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:32.691502 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:32.791931 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:32.892266 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:32.992448 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:33.092741 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:33.193232 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:33.293564 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:33.295376 2927 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:33.393609 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:33.419795 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:26:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:33.421000 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:26:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:33.421039 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:26:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:33.421054 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:26:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:33.421077 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:26:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:33.421870 2927 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:26:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:33.494124 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:33.594792 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:33.614762 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:33.695174 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:33.795656 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:33.896152 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:33.997214 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:34.097521 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:34.197851 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:34.298182 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:34.398672 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:34.499179 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:34.599782 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:34.614812 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:34.700801 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:34.801305 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:34.901744 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:35.002352 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:35.102697 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:35.203194 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:35.303639 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:35.404121 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:35.504607 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:35.605184 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:35.614201 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:35.705558 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:35.806158 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:35.906404 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:36.006936 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:36.107216 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:36.207533 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:36.307931 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:36.408366 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:36.508746 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:36.609259 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:36.614094 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:36.709606 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:36.809928 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:36.910120 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:37.010860 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:37.111133 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:37.211658 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:37.311991 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:37.412442 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:37.512795 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:26:37.514601 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:37.514648 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:37.613058 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:37.613840 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:37.713422 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:37.813821 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:37.914014 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:38.014699 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:38.114930 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:38.215240 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:38.315740 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:38.415796 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:38.515932 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:38.614494 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:38.616572 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:38.717288 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:38.817795 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:38.918298 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:38.937450 2927 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:39.019044 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:39.119380 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:39.219891 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:39.320347 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:39.420858 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:39.521116 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:39.614604 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:39.621678 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:39.722297 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:39.822794 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:39.923181 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:40.023896 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:40.124236 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:40.224733 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:40.296625 2927 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:40.324993 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:40.422358 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:26:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:40.423502 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:26:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:40.423626 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:26:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:40.423643 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:26:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:40.423667 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:26:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:40.424426 2927 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:26:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:40.425452 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:40.525930 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:40.613993 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:40.626019 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:40.726740 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:40.757558 2927 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.173779953e5bec41", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientMemory", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientMemory", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 494728257, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 494728257, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:26:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:40.827008 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:40.927513 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:41.028109 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:41.128505 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:41.228826 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:41.329201 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:41.429697 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:41.530046 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:41.614253 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:41.630333 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:41.731045 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:41.831303 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:41.931793 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:42.032543 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:42.132772 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:42.233225 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:42.333578 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:42.434051 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:42.534587 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:42.614776 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:42.634857 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:42.735463 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:42.835851 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:42.936055 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:43.036580 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:43.136911 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:43.237218 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:43.337659 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:43.438101 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:43.538670 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:43.613788 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:43.638863 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:43.739262 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:43.839591 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:43.939928 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:44.040864 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:44.141053 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:44.241230 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:44.341659 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:26:44.441802633Z" level=error msg="Failed to cleanup (probably retrying): failed to destroy network for pod sandbox k8s_thanos-querier-6f4f5fb7cf-rdm9l_openshift-monitoring_7c63c9be-d1ad-44b6-8b33-09af3ee314af_0(f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3): error removing pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/7c63c9be-d1ad-44b6-8b33-09af3ee314af]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" Jan 05 09:26:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:44.441947 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:44.542815 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:44.614067 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:44.643274 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:44.743997 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:44.844347 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:44.944932 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:45.045501 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:45.146394 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:45.246887 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:45.347373 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:45.447712 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:45.548293 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:26:45.567739343Z" level=info msg="cleanup sandbox network" Jan 05 09:26:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:26:45.568081012Z" level=info msg="Got pod network &{Name:thanos-querier-6f4f5fb7cf-rdm9l Namespace:openshift-monitoring ID:f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3 UID:7c63c9be-d1ad-44b6-8b33-09af3ee314af NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:26:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:26:45.568274256Z" level=info msg="Deleting pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:26:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:45.614053 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:45.649214 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:45.749932 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:45.850417 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:45.950998 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:46.051562 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:46.151734 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:46.252177 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:46.352655 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:46.453144 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:46.553771 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:46.614812 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:46.654009 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:46.754477 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:46.854931 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:46.955541 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:47.056064 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:47.156294 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:47.256724 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:47.297585 2927 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:47.357016 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:47.425385 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:26:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:47.426534 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:26:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:47.426564 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:26:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:47.426577 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:26:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:47.426602 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:26:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:47.427410 2927 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:26:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:47.457557 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:47.558156 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:47.613942 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:47.659170 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:47.759715 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:47.860162 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:47.960599 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:48.061254 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:48.161490 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:48.261938 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:48.362424 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:48.462929 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:48.563108 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:48.614280 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:48.663602 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:48.764169 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:48.864640 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:48.938129 2927 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:48.965160 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:49.065819 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:49.165997 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:49.266478 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:49.366936 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:49.467277 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:49.567560 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:49.614583 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:49.667896 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:49.768487 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:49.868928 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:49.969503 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:50.070195 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:50.170360 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:50.270722 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:50.371160 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:50.471508 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:50.572116 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:50.614413 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:50.672602 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:50.758895 2927 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.173779953e5bec41", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientMemory", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientMemory", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 494728257, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 494728257, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:26:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:50.772990 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:50.873464 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:50.973913 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:51.074289 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:51.174508 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:51.274993 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:51.375461 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:51.475715 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:51.575990 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:51.614023 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:51.676337 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:51.776931 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:51.877400 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:51.978004 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:52.078671 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:52.178871 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:52.279164 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:52.379631 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:52.480113 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:52.580660 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:52.614660 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:52.680934 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:52.781476 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:52.881935 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:52.982518 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:53.083051 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:53.183256 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:53.284275 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:53.385347 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:53.485581 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:53.586301 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:53.614263 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:53.686582 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:53.787167 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:53.887637 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:53.987690 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:54.088361 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:54.188466 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:54.288938 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:54.298802 2927 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:54.389298 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:54.428257 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:26:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:54.429277 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:26:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:54.429316 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:26:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:54.429333 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:26:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:54.429354 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:26:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:54.430080 2927 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:26:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:54.490291 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:54.590846 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:54.613759 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:54.691055 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:54.791479 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:54.891933 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:54.992539 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:55.093000 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:55.193181 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:55.293514 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:55.393992 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:55.494499 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:55.595057 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:55.614005 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:55.695227 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:55.795436 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:55.895919 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:55.996496 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:56.097177 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:56.197367 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:56.297851 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:56.398293 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:56.498734 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:56.599301 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:56.614242 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:56.699692 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:56.799745 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:56.900218 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:57.000835 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:57.101486 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:57.201682 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:57.302158 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:57.402575 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:57.502907 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:57.603506 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:57.614466 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:57.704519 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:57.805002 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:57.905481 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:58.005826 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:58.106442 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:58.207438 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:58.307762 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:58.408177 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:58.508752 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:26:58.557078 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:58.557130 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:58.609281 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:58.614076 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:58.709584 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:58.810148 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:58.910629 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:58.939003 2927 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:59.011054 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:59.111325 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:59.211645 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:59.311938 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:59.412317 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:59.512587 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:59.612772 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:26:59.614556 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:26:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:59.713118 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:59.813572 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:26:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:26:59.913933 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:00.014515 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:00.114701 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:00.215118 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:00.315565 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:00.416041 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:00.516116 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:00.614352 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:00.616410 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:00.717123 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:00.760023 2927 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.173779953e5bec41", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientMemory", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientMemory", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 494728257, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 494728257, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:27:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:00.817469 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:00.917988 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:01.018662 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:01.118724 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:01.219171 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:01.300028 2927 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:01.319341 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:01.419702 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:01.430618 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:27:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:01.431762 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:27:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:01.431879 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:27:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:01.431897 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:27:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:01.431935 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:27:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:01.432713 2927 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:27:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:01.520053 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:01.614204 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:01.620286 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:01.720935 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:01.821370 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:01.921740 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:02.022352 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:02.122533 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:02.223022 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:02.323530 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:02.424021 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:02.524604 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:02.614782 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:02.624863 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:02.725351 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:02.825700 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:02.926172 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:03.026925 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:03.127266 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:03.227767 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:03.328344 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:03.428852 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:03.529485 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:03.614563 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:03.629598 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:03.730277 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:03.830597 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:03.931050 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:04.031283 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:04.131365 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:04.231858 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:04.332358 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:04.432842 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:04.533145 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:04.614434 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:04.633512 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:04.734171 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:04.834613 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:04.934994 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:05.035641 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:05.135835 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:05.236251 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:05.336715 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:05.437168 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:05.537670 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:05.614792 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:05.637873 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:05.738556 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:05.838881 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:05.939442 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:06.039932 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:06.140925 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:06.241140 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:06.341663 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:06.441891 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:06.542326 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:06.614457 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:06.642598 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:06.743288 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:06.843740 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:06.944173 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:07.044677 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:07.144856 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:07.245360 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:07.345740 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:07.446163 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:07.546674 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:07.614679 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:07.646881 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:07.747624 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:07.847934 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:07.948541 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:08.049221 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:08.150259 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:08.250629 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:08.301475 2927 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:08.350913 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:27:08.400796 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:08.400847 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:08.433216 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:27:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:08.434645 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:27:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:08.434789 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:27:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:08.434862 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:27:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:08.434954 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:27:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:08.435699 2927 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:27:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:08.451780 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:08.552845 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:08.613952 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:08.653115 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:08.753832 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:08.854283 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:08.939842 2927 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:08.954442 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:09.055476 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:09.155704 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:09.256168 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:09.356644 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:09.457107 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:09.557704 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:09.614505 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:09.658651 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:09.759218 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:09.859311 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:09.959836 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:10.060535 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:10.160803 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:10.261116 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:10.361583 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:10.462050 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:10.562596 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:10.614767 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:10.662936 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:10.761417 2927 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.173779953e5bec41", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientMemory", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientMemory", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 494728257, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 494728257, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:27:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:10.763484 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:10.863933 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:10.964529 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:11.065111 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:11.165336 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:11.265662 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:11.366110 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:11.466601 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:11.567052 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:11.613761 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:11.667936 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:11.768643 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:11.869118 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:11.969695 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:12.070411 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:12.171440 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:12.271930 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:12.372408 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:12.472878 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:12.573479 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:12.614447 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:12.673768 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:12.774368 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:12.874839 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:12.975432 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:13.075498 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:13.176278 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:13.276718 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:13.377165 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:13.477550 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:27:13.525769 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:13.525821 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:13.578053 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:13.614054 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:13.678372 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:13.778833 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:13.879296 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:13.980293 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:14.080761 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:14.180935 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:14.281898 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:14.382361 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:14.483427 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:14.583701 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:14.614640 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:14.683936 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:14.784484 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:14.884928 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:14.985531 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:15.086296 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:15.187307 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:15.288497 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:15.302350 2927 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:15.389554 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:15.435929 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:27:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:15.437011 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:27:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:15.437041 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:27:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:15.437054 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:27:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:15.437077 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:27:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:15.437827 2927 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:27:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:15.489836 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:15.590411 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:15.614333 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:15.690698 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:15.791126 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:15.891603 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:15.992166 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:16.092735 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:16.193371 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:16.293852 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:16.394294 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:16.494500 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:16.595055 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:16.614074 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:16.695388 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:16.795939 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:16.896159 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:16.996235 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:17.096898 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:17.197131 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:17.297607 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:17.398058 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:17.498489 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:17.599055 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:17.614009 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:17.699336 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:17.800184 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:17.900767 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:18.001346 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:18.102047 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:18.202990 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:18.303467 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:18.403935 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:18.504572 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:18.605405 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:18.614300 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:18.705701 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:18.806176 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:18.906527 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:18.939994 2927 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:19.007051 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:19.107685 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:19.208752 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:19.309113 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:19.409547 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:19.509936 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:19.610475 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:19.614487 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:19.710877 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:19.811233 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:19.911728 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:20.012287 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:20.112917 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:20.213820 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:20.314033 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:20.414231 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:20.514670 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:20.614205 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:20.615225 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:20.715767 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:20.762616 2927 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.173779953e5bec41", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientMemory", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientMemory", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 494728257, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 494728257, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:27:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:20.815868 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:20.916360 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:21.016991 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:21.117747 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:21.218385 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:21.318900 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:21.419349 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:21.519720 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:21.613970 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:21.619994 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:21.720745 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:21.821056 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:21.921534 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:22.021935 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:22.122616 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:22.223055 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:22.304071 2927 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:22.323378 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:22.423638 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:22.438761 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:27:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:22.439995 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:27:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:22.440025 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:27:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:22.440061 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:27:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:22.440085 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:27:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:22.440957 2927 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:27:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:22.524389 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:22.614508 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:22.624576 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:22.725291 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:22.825791 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:22.926273 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:23.026933 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:23.127546 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:23.228555 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:23.328604 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:23.429368 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:23.529859 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:23.614633 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:23.630725 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:23.731356 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:23.831930 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:23.932347 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:24.032516 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:24.133059 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:24.233265 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:24.333752 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:24.434243 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:24.534732 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:24.614754 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:24.634783 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:24.735299 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:24.835800 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:24.936021 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:25.036697 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:25.137302 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:25.237512 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:25.337996 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:25.438347 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:25.538689 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:25.614709 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:25.638739 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:25.739169 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:25.839683 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:25.940285 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:26.040884 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:26.140948 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:26.241763 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:26.342221 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:26.442668 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:26.543106 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:26.614248 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:26.643334 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:26.743831 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:26.844178 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:26.944589 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:27.045144 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:27.146194 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:27.246627 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:27.347101 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:27.447594 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:27.548171 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:27.614201 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:27.648237 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:27.748593 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:27.848931 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:27.949257 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:28.049911 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:28.150084 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:28.250377 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:28.350877 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:28.451360 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:28.552237 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:28.614647 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:28.652842 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:28.753269 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:28.853863 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:28.940900 2927 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:28.953924 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:29.054113 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:29.154463 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:29.254932 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:29.305750 2927 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:29.355229 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:29.441709 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:27:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:29.442873 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:27:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:29.442914 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:27:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:29.442929 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:27:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:29.442951 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:27:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:29.443676 2927 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:27:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:29.455756 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:29.556355 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:29.614297 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:29.657277 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:29.758003 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:29.858477 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:29.958932 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:30.059666 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:30.159911 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:30.260378 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:30.360722 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:30.461001 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:30.561557 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:27:30.578816248Z" level=error msg="Failed to cleanup (probably retrying): failed to destroy network for pod sandbox k8s_thanos-querier-6f4f5fb7cf-rdm9l_openshift-monitoring_7c63c9be-d1ad-44b6-8b33-09af3ee314af_0(f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3): error removing pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/7c63c9be-d1ad-44b6-8b33-09af3ee314af]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" Jan 05 09:27:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:30.614562 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:30.661741 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:30.762405 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:30.764187 2927 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.173779953e5bec41", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientMemory", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientMemory", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 494728257, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 17, 24, 28, 494728257, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:27:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:30.862649 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:30.963242 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:31.063847 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:31.164867 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:31.265345 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:31.365732 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:31.466161 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:31.566772 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:31.614749 2927 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:31.666933 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:31.767643 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:31.867935 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:27:31.950288 2927 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:31.950337 2927 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:27:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:31.968355 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:32.068989 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:32.169142 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:27:32.267246371Z" level=info msg="cleanup sandbox network" Jan 05 09:27:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:27:32.267481753Z" level=info msg="Got pod network &{Name:thanos-querier-6f4f5fb7cf-rdm9l Namespace:openshift-monitoring ID:f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3 UID:7c63c9be-d1ad-44b6-8b33-09af3ee314af NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:27:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:27:32.267645114Z" level=info msg="Deleting pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:27:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:32.269719 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:32.370479 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:32.471049 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:32.571489 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:32.672242 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:32.772898 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:32.873167 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:32.973702 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:33.074414 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:33.174671 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:33.275146 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:33.375625 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:33.476114 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:33.576715 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:33.677185 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:33.777813 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:33.878160 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:33.979260 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:34.079806 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:34.180027 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:34.280516 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:34.381051 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:34.481469 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:34.582053 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:34.682487 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:34.783120 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:34.883406 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:34.983995 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:35.084517 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:35.184797 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:35.285354 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:35.385653 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:35.486121 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:35.586365 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:35.686937 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:35.787159 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:35.887557 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:35.988011 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:36.088705 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:36.189750 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:36.290059 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:36.391106 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:36.444476 2927 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:36.445612 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:36.445645 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:36.445661 2927 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:36.445681 2927 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:36.461307 2927 kubelet_node_status.go:110] "Node was previously registered" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:27:36.461491 2927 kubelet_node_status.go:75] "Successfully registered node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:36.491827 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:36.592228 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:36.692825 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:36.793352 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:36.893706 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:36.994035 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:37.094455 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:37.194670 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:37.295177 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:37.395665 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:37.495992 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:37.596605 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:37.697182 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:37.797682 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:37.898156 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:37.998258 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:38.098925 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:38.199000 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:38.299600 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:38.400044 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:38.500597 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:38.601301 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:38.701858 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:38.802296 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:38.902719 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:38.941190 2927 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:39.003185 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:39.103894 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:39.204847 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:39.305300 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:39.405719 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:39.506161 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:39.606693 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:39.707298 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:39.807619 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:39.907820 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:40.008354 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:40.108644 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:40.209614 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:40.310000 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:40.410421 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:40.510934 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:40.611382 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:40.711990 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:40.812157 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:40.913070 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:41.013616 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:41.114204 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:41.214527 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:41.315000 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:41.415992 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:41.516504 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:41.616930 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:41.717675 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:41.818138 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:41.918375 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:42.019054 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:42.119765 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:42.220388 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:42.320843 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:42.421396 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:42.521916 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:42.622942 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:42.723656 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:42.824496 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:42.924746 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:43.025054 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:43.125635 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:43.225995 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:43.326477 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:43.427236 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:43.527455 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:43.628232 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:43.728700 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:43.829237 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:43.929616 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:44.030698 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:44.131351 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:44.231570 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:44.332053 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:44.432545 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:44.532935 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:44.633106 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:44.733712 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:44.834168 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:44.934507 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:45.035147 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:45.135945 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:45.236163 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:45.336592 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:45.437052 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:45.537392 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:45.637843 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:45.738407 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:45.838764 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:45.939234 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:46.039844 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:46.140376 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:46.240592 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:46.341045 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:46.441538 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:46.541598 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:46.642006 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:46.742720 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:46.843576 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:46.944176 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:47.044736 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:47.145773 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:47.246665 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:47.347042 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:47.447306 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:47.547704 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:47.647941 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:47.748601 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:47.849046 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:47.949654 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:48.050342 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:48.150858 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:48.250934 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:48.351365 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:48.451743 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:48.552356 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:48.652927 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:48.753600 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:48.854106 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:48.941638 2927 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:48.955170 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:49.055211 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:49.155553 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:49.255656 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:49.356134 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:49.456676 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:49.557130 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:49.657604 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:49.758285 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:49.858708 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:49.958932 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:50.059655 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:50.160140 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:50.261163 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:50.361482 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:50.461933 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:50.562529 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:50.663063 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:50.763779 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:50.864250 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:50.964802 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:51.065667 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:51.165934 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:51.266119 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:51.366594 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:51.467055 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:51.567565 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:51.667999 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:51.768577 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:51.868934 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:51.969524 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:52.070255 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:52.170581 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:52.271694 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:52.372160 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:52.472646 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:52.573237 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:52.673717 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:52.774409 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:52.874900 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:52.975425 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:53.075939 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:53.176933 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:53.277529 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:53.378051 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:53.478382 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:53.578831 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:53.679245 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:53.779931 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:53.880391 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:53.980576 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:54.081109 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:54.181349 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:54.281669 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:54.382146 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:54.482618 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:54.583114 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:54.683633 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:54.784306 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:54.884740 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:54.984997 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:55.085576 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:55.186672 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:55.286998 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:55.387471 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:55.487802 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:55.588418 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:55.688935 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:55.789598 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:55.889900 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:55.990515 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:56.090931 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:56.191168 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:56.291666 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:56.391993 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:56.492507 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:56.593067 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:56.693515 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:56.794480 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:56.894782 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:56.995181 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:57.095852 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:57.196029 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:57.296083 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:57.396529 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:57.496994 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:57.597604 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:57.698120 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:57.798702 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:57.899013 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:57.999616 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:58.100286 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:58.200363 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:58.300858 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:58.401284 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:58.501861 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:58.602347 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:58.702936 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:58.803528 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:58.904239 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:58.942663 2927 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:59.004709 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:59.105356 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:59.205585 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:59.306060 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:59.406592 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:59.507114 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:59.607348 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:59.707939 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:59.808408 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:27:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:27:59.908765 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:00.009247 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:00.109855 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:00.210070 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:00.310560 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:00.410987 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:00.511328 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:00.611687 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:00.712298 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:00.812833 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:00.913261 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:01.013822 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:01.114491 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:01.215564 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:01.316056 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:01.416534 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:01.516990 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:01.617581 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:01.718170 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:01.818519 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:01.919014 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:02.019626 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:02.120290 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:02.221114 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:02.321603 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:02.422114 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:02.522602 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:02.623050 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:02.723605 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:02.824170 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:02.924603 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:03.025176 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:03.125885 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:03.226894 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:03.327384 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:03.427861 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:03.528487 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:03.629054 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:03.729461 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:03.829996 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:03.930281 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:04.031297 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:04.131999 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:04.232253 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:04.332750 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:28:04.433113 2927 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.701446 2927 apiserver.go:52] "Watching apiserver" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.709592 2927 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-monitoring/telemeter-client-5598c645c4-mftdv openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4 openshift-dns/node-resolver-nxzr8 openshift-multus/multus-pbfzz openshift-network-diagnostics/network-check-target-xmq2g openshift-monitoring/prometheus-k8s-0 openshift-cluster-node-tuning-operator/tuned-smwj7 openshift-dns/dns-default-2pfzf openshift-machine-config-operator/machine-config-daemon-hf8f5 openshift-multus/multus-additional-cni-plugins-5fgxc openshift-ingress/router-default-5cf6bbdff9-tnffs openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256 openshift-ovn-kubernetes/ovnkube-node-rp9bd openshift-monitoring/node-exporter-2r6nf openshift-ingress-canary/ingress-canary-8xrbh openshift-image-registry/node-ca-xw6d2 openshift-monitoring/kube-state-metrics-6798b548b9-xl88q openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l openshift-multus/network-metrics-daemon-hq594 openshift-monitoring/alertmanager-main-0 openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh] Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.709694 2927 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.709806 2927 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.709873 2927 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.710750 2927 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.712249 2927 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.712421 2927 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.712567 2927 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.712741 2927 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.712874 2927 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.713042 2927 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.713179 2927 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.713339 2927 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.713494 2927 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.713671 2927 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.713894 2927 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.714240 2927 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.715505 2927 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.715739 2927 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.716034 2927 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.716324 2927 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.718300 2927 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.718470 2927 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.794171 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8f88cbe5-1221-4e6f-a6c9-f22da152b43f-hosts-file\") pod \"node-resolver-nxzr8\" (UID: \"8f88cbe5-1221-4e6f-a6c9-f22da152b43f\") " pod="openshift-dns/node-resolver-nxzr8" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.794359 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-volume-directive-shadow\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.794478 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-lib-modules\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.794590 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-cni-bin\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.794849 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wq4k7\" (UniqueName: \"kubernetes.io/projected/c0e47249-a5ff-460e-958d-53577387c154-kube-api-access-wq4k7\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.794992 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-cni-binary-copy\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.795096 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-root\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.795197 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-config\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.795311 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-openshift-state-metrics-tls\") pod \"openshift-state-metrics-6f5dcdf79b-wq256\" (UID: \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\") " pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.795563 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-multus-cni-dir\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.795682 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.795785 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-sys\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.795874 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-cnibin\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.795979 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-metrics-client-ca\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.796224 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-proxy\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-main-proxy\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.796337 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.796430 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-metrics-certs\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.796522 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/a40bf20d-67a4-45e7-b701-ca5a75376c1f-audit-log\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.796703 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-run-ovn-kubernetes\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.796803 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-ca\" (UniqueName: \"kubernetes.io/configmap/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovn-ca\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.796986 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3b10ff4-a243-4a4e-b27d-f613e258666d-service-ca-bundle\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.797103 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfs6g\" (UniqueName: \"kubernetes.io/projected/a40bf20d-67a4-45e7-b701-ca5a75376c1f-kube-api-access-gfs6g\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.797213 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/05c9a6ea-140a-49a2-bde5-f220b05aa252-tls-assets\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.797345 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-env-overrides\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.797444 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-system-cni-dir\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.797756 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzzxr\" (UniqueName: \"kubernetes.io/projected/1bb84a70-232c-4467-ac0b-647030155908-kube-api-access-vzzxr\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.797885 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.798012 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.798108 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.798215 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-state-metrics-tls\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.798296 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-etc\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.798560 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfs2p\" (UniqueName: \"kubernetes.io/projected/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-kube-api-access-dfs2p\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.798681 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.798804 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-kubelet-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.798925 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-registration-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.799028 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-system-cni-dir\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.799122 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls\" (UniqueName: \"kubernetes.io/secret/a40bf20d-67a4-45e7-b701-ca5a75376c1f-tls\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.799207 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/05c9a6ea-140a-49a2-bde5-f220b05aa252-config-out\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.799309 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-systemd-units\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.799398 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovn-node-metrics-cert\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.799487 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx2gt\" (UniqueName: \"kubernetes.io/projected/50152296-9489-4fa8-aa42-7858debe1c08-kube-api-access-zx2gt\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.799584 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-metrics-client-ca\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.799685 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-default-certificate\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.799775 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-config-out\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.799855 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-web-config\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.799954 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-os-release\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.800057 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/7c63c9be-d1ad-44b6-8b33-09af3ee314af-metrics-client-ca\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.800145 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvz7l\" (UniqueName: \"kubernetes.io/projected/7c63c9be-d1ad-44b6-8b33-09af3ee314af-kube-api-access-pvz7l\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.800229 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-run-openvswitch\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.800312 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-sys\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.800395 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.800482 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7xz4\" (UniqueName: \"kubernetes.io/projected/91265f01-9dcb-4046-961e-ad4c544224d7-kube-api-access-b7xz4\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.800568 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/05c9a6ea-140a-49a2-bde5-f220b05aa252-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.800674 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxsm5\" (UniqueName: \"kubernetes.io/projected/05c9a6ea-140a-49a2-bde5-f220b05aa252-kube-api-access-kxsm5\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.800764 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.800846 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-querier-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c63c9be-d1ad-44b6-8b33-09af3ee314af-thanos-querier-trusted-ca-bundle\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.803925 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccjjw\" (UniqueName: \"kubernetes.io/projected/c3b10ff4-a243-4a4e-b27d-f613e258666d-kube-api-access-ccjjw\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.803980 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/659b8bb4-81b5-49bf-8127-1d326f30ba2f-serviceca\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.804019 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/88745ab7-efe7-42c0-a2e8-75591518333d-tls-certificates\") pod \"prometheus-operator-admission-webhook-598468bb49-94dsn\" (UID: \"88745ab7-efe7-42c0-a2e8-75591518333d\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.804052 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-wtmp\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.804085 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-config-volume\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.804119 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-run-ovn\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.804151 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovnkube-config\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.804182 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-metrics-tls\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.804218 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmxtc\" (UniqueName: \"kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc\") pod \"network-check-target-xmq2g\" (UID: \"f364a949-09d7-4cfb-83ff-e532b822a557\") " pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.804250 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7www\" (UniqueName: \"kubernetes.io/projected/8f88cbe5-1221-4e6f-a6c9-f22da152b43f-kube-api-access-g7www\") pod \"node-resolver-nxzr8\" (UID: \"8f88cbe5-1221-4e6f-a6c9-f22da152b43f\") " pod="openshift-dns/node-resolver-nxzr8" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.804281 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd-system\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-run-systemd-system\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.804316 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-tuned-profiles-data\" (UniqueName: \"kubernetes.io/configmap/91265f01-9dcb-4046-961e-ad4c544224d7-var-lib-tuned-profiles-data\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.804356 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-web-config\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.804390 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xrks\" (UniqueName: \"kubernetes.io/projected/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-kube-api-access-5xrks\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.804441 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/659b8bb4-81b5-49bf-8127-1d326f30ba2f-host\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.804486 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-run-netns\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.804522 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-os-release\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.804556 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-cert\" (UniqueName: \"kubernetes.io/secret/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovn-cert\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.804591 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-kube-etcd-client-certs\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-kube-etcd-client-certs\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.804626 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.804681 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.804715 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98gp9\" (UniqueName: \"kubernetes.io/projected/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-kube-api-access-98gp9\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.804749 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-grpc-tls\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.804783 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/05c9a6ea-140a-49a2-bde5-f220b05aa252-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.804821 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.804858 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c0e47249-a5ff-460e-958d-53577387c154-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.804893 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a40bf20d-67a4-45e7-b701-ca5a75376c1f-tmpfs\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.804946 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-etc-openvswitch\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.804986 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cookie-secret\" (UniqueName: \"kubernetes.io/secret/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-cookie-secret\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.805020 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-plugin-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.805055 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-cnibin\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.805087 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-node-log\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.805121 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-cni-netd\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.805154 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.805190 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-host\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.805224 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-adapter-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-prometheus-adapter-audit-profiles\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.805260 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.805295 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c0e47249-a5ff-460e-958d-53577387c154-cni-binary-copy\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.805328 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-stats-auth\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.805362 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-device-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.805399 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-serving-certs-ca-bundle\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.805435 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemeter-client-tls\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-telemeter-client-tls\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.805472 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r64p\" (UniqueName: \"kubernetes.io/projected/fc45e847-8f55-4c48-8c48-4bfee2419dca-kube-api-access-4r64p\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.805508 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.805547 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.805584 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blpwr\" (UniqueName: \"kubernetes.io/projected/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-api-access-blpwr\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.805622 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.805677 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.805711 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.805748 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4sjn\" (UniqueName: \"kubernetes.io/projected/659b8bb4-81b5-49bf-8127-1d326f30ba2f-kube-api-access-h4sjn\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.805784 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g48vj\" (UniqueName: \"kubernetes.io/projected/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-kube-api-access-g48vj\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.805822 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-oauth-cookie\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-oauth-cookie\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.805859 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.805896 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05c9a6ea-140a-49a2-bde5-f220b05aa252-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.805957 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-telemeter-client-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-secret-telemeter-client-kube-rbac-proxy-config\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.805995 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.806033 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemeter-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-telemeter-trusted-ca-bundle\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.806068 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tv6rn\" (UniqueName: \"kubernetes.io/projected/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-kube-api-access-tv6rn\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.806107 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.806142 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csks4\" (UniqueName: \"kubernetes.io/projected/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-kube-api-access-csks4\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.806198 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6g5r\" (UniqueName: \"kubernetes.io/projected/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-kube-api-access-h6g5r\") pod \"openshift-state-metrics-6f5dcdf79b-wq256\" (UID: \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\") " pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.806237 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-var-lib-openvswitch\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.806271 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-proxy\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-proxy\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.806306 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-tls\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.806345 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"non-standard-root-system-trust-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bb84a70-232c-4467-ac0b-647030155908-non-standard-root-system-trust-ca-bundle\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.806381 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-telemeter-client\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-secret-telemeter-client\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.806424 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-config\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.806463 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-tls\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.806498 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-var-run-dbus\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.806535 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-metrics-client-ca\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.806573 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-6f5dcdf79b-wq256\" (UID: \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\") " pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.806614 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-serving-certs-ca-bundle\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.806667 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-log-socket\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.806703 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzfn5\" (UniqueName: \"kubernetes.io/projected/e04a47b4-aa8e-4d4e-9d45-aa8037dcb748-kube-api-access-lzfn5\") pod \"ingress-canary-8xrbh\" (UID: \"e04a47b4-aa8e-4d4e-9d45-aa8037dcb748\") " pod="openshift-ingress-canary/ingress-canary-8xrbh" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.806742 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.806777 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-proxy-tls\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.806811 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.806847 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-config-volume\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.806880 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-textfile\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.806934 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-adapter-prometheus-config\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-prometheus-adapter-prometheus-config\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.806983 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.807019 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-rootfs\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.807055 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-slash\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.807069 2927 reconciler.go:169] "Reconciler: start to sync state" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.919357 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-cni-binary-copy\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.919418 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-root\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.919454 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-config\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.919503 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-openshift-state-metrics-tls\") pod \"openshift-state-metrics-6f5dcdf79b-wq256\" (UID: \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\") " pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.919553 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-multus-cni-dir\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.919584 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.919625 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-sys\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.919672 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-cnibin\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.919717 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-metrics-client-ca\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.919749 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-proxy\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-main-proxy\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.919794 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.919825 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-metrics-certs\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.919867 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/a40bf20d-67a4-45e7-b701-ca5a75376c1f-audit-log\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.919900 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-run-ovn-kubernetes\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.919949 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"ovn-ca\" (UniqueName: \"kubernetes.io/configmap/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovn-ca\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.919995 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3b10ff4-a243-4a4e-b27d-f613e258666d-service-ca-bundle\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.920029 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-gfs6g\" (UniqueName: \"kubernetes.io/projected/a40bf20d-67a4-45e7-b701-ca5a75376c1f-kube-api-access-gfs6g\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.920059 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/05c9a6ea-140a-49a2-bde5-f220b05aa252-tls-assets\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.920105 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-env-overrides\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.920139 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-system-cni-dir\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.920186 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-vzzxr\" (UniqueName: \"kubernetes.io/projected/1bb84a70-232c-4467-ac0b-647030155908-kube-api-access-vzzxr\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.920222 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.920273 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.920321 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.920357 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-state-metrics-tls\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.920482 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"etc\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-etc\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.920531 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-dfs2p\" (UniqueName: \"kubernetes.io/projected/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-kube-api-access-dfs2p\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.920566 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.920611 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-kubelet-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.920645 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-registration-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.920710 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-system-cni-dir\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.920754 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tls\" (UniqueName: \"kubernetes.io/secret/a40bf20d-67a4-45e7-b701-ca5a75376c1f-tls\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.920787 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/05c9a6ea-140a-49a2-bde5-f220b05aa252-config-out\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.920818 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-systemd-units\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.920867 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovn-node-metrics-cert\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.920925 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-zx2gt\" (UniqueName: \"kubernetes.io/projected/50152296-9489-4fa8-aa42-7858debe1c08-kube-api-access-zx2gt\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.920963 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-metrics-client-ca\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.921009 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-default-certificate\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.921042 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-config-out\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.921086 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-web-config\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.921118 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-os-release\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.921167 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/7c63c9be-d1ad-44b6-8b33-09af3ee314af-metrics-client-ca\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.921205 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-pvz7l\" (UniqueName: \"kubernetes.io/projected/7c63c9be-d1ad-44b6-8b33-09af3ee314af-kube-api-access-pvz7l\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.921249 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-run-openvswitch\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.921282 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-sys\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.921329 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.921363 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-b7xz4\" (UniqueName: \"kubernetes.io/projected/91265f01-9dcb-4046-961e-ad4c544224d7-kube-api-access-b7xz4\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.921408 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/05c9a6ea-140a-49a2-bde5-f220b05aa252-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.921442 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-kxsm5\" (UniqueName: \"kubernetes.io/projected/05c9a6ea-140a-49a2-bde5-f220b05aa252-kube-api-access-kxsm5\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.921489 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.921526 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"thanos-querier-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c63c9be-d1ad-44b6-8b33-09af3ee314af-thanos-querier-trusted-ca-bundle\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.921573 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-ccjjw\" (UniqueName: \"kubernetes.io/projected/c3b10ff4-a243-4a4e-b27d-f613e258666d-kube-api-access-ccjjw\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.921604 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/659b8bb4-81b5-49bf-8127-1d326f30ba2f-serviceca\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.921671 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/88745ab7-efe7-42c0-a2e8-75591518333d-tls-certificates\") pod \"prometheus-operator-admission-webhook-598468bb49-94dsn\" (UID: \"88745ab7-efe7-42c0-a2e8-75591518333d\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.921718 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-wtmp\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.921752 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-config-volume\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.921799 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-run-ovn\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.921832 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovnkube-config\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.921875 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-metrics-tls\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.921931 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-mmxtc\" (UniqueName: \"kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc\") pod \"network-check-target-xmq2g\" (UID: \"f364a949-09d7-4cfb-83ff-e532b822a557\") " pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.921965 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-g7www\" (UniqueName: \"kubernetes.io/projected/8f88cbe5-1221-4e6f-a6c9-f22da152b43f-kube-api-access-g7www\") pod \"node-resolver-nxzr8\" (UID: \"8f88cbe5-1221-4e6f-a6c9-f22da152b43f\") " pod="openshift-dns/node-resolver-nxzr8" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.922009 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"run-systemd-system\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-run-systemd-system\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.922045 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"var-lib-tuned-profiles-data\" (UniqueName: \"kubernetes.io/configmap/91265f01-9dcb-4046-961e-ad4c544224d7-var-lib-tuned-profiles-data\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.922089 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-web-config\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.922124 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-5xrks\" (UniqueName: \"kubernetes.io/projected/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-kube-api-access-5xrks\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.922167 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/659b8bb4-81b5-49bf-8127-1d326f30ba2f-host\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.922202 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-run-netns\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.922248 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-os-release\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.922279 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"ovn-cert\" (UniqueName: \"kubernetes.io/secret/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovn-cert\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.922325 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-kube-etcd-client-certs\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-kube-etcd-client-certs\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.922359 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.922405 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.922438 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-98gp9\" (UniqueName: \"kubernetes.io/projected/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-kube-api-access-98gp9\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.922488 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-grpc-tls\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.922524 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/05c9a6ea-140a-49a2-bde5-f220b05aa252-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.922572 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.922608 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c0e47249-a5ff-460e-958d-53577387c154-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.922672 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a40bf20d-67a4-45e7-b701-ca5a75376c1f-tmpfs\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.922721 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-etc-openvswitch\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.922756 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cookie-secret\" (UniqueName: \"kubernetes.io/secret/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-cookie-secret\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.922804 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"plugin-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-plugin-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.922838 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-cnibin\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.922893 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-node-log\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.922949 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-cni-netd\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.922983 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.923026 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-host\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.923075 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"prometheus-adapter-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-prometheus-adapter-audit-profiles\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.923112 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.923147 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c0e47249-a5ff-460e-958d-53577387c154-cni-binary-copy\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.923192 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-stats-auth\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.923230 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-device-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.923266 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-serving-certs-ca-bundle\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.923314 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"telemeter-client-tls\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-telemeter-client-tls\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.923354 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-4r64p\" (UniqueName: \"kubernetes.io/projected/fc45e847-8f55-4c48-8c48-4bfee2419dca-kube-api-access-4r64p\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.923391 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.923432 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.923471 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-blpwr\" (UniqueName: \"kubernetes.io/projected/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-api-access-blpwr\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.923506 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.923563 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.923603 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.923639 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-h4sjn\" (UniqueName: \"kubernetes.io/projected/659b8bb4-81b5-49bf-8127-1d326f30ba2f-kube-api-access-h4sjn\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.923687 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-g48vj\" (UniqueName: \"kubernetes.io/projected/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-kube-api-access-g48vj\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.923726 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-oauth-cookie\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-oauth-cookie\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.923761 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.923798 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05c9a6ea-140a-49a2-bde5-f220b05aa252-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.923837 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-telemeter-client-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-secret-telemeter-client-kube-rbac-proxy-config\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.924371 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.924420 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"telemeter-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-telemeter-trusted-ca-bundle\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.924455 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-tv6rn\" (UniqueName: \"kubernetes.io/projected/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-kube-api-access-tv6rn\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.924500 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.924536 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-csks4\" (UniqueName: \"kubernetes.io/projected/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-kube-api-access-csks4\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.924572 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-h6g5r\" (UniqueName: \"kubernetes.io/projected/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-kube-api-access-h6g5r\") pod \"openshift-state-metrics-6f5dcdf79b-wq256\" (UID: \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\") " pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.924607 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-var-lib-openvswitch\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.924680 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-proxy\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-proxy\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.924716 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-tls\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.924761 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"non-standard-root-system-trust-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bb84a70-232c-4467-ac0b-647030155908-non-standard-root-system-trust-ca-bundle\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.924801 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-telemeter-client\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-secret-telemeter-client\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.924835 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-config\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.924871 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-tls\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.924914 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-var-run-dbus\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.924954 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-metrics-client-ca\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.924994 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-6f5dcdf79b-wq256\" (UID: \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\") " pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.925028 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-serving-certs-ca-bundle\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.925063 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-log-socket\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.925098 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-lzfn5\" (UniqueName: \"kubernetes.io/projected/e04a47b4-aa8e-4d4e-9d45-aa8037dcb748-kube-api-access-lzfn5\") pod \"ingress-canary-8xrbh\" (UID: \"e04a47b4-aa8e-4d4e-9d45-aa8037dcb748\") " pod="openshift-ingress-canary/ingress-canary-8xrbh" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.925133 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.925167 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-proxy-tls\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.925201 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.925236 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-config-volume\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.925271 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-textfile\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.925311 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"prometheus-adapter-prometheus-config\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-prometheus-adapter-prometheus-config\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.925346 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.925380 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-rootfs\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.925413 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-slash\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.925445 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8f88cbe5-1221-4e6f-a6c9-f22da152b43f-hosts-file\") pod \"node-resolver-nxzr8\" (UID: \"8f88cbe5-1221-4e6f-a6c9-f22da152b43f\") " pod="openshift-dns/node-resolver-nxzr8" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.925486 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-volume-directive-shadow\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.925519 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-lib-modules\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.925557 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-cni-bin\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.925594 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-wq4k7\" (UniqueName: \"kubernetes.io/projected/c0e47249-a5ff-460e-958d-53577387c154-kube-api-access-wq4k7\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.929316 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-cni-binary-copy\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.929426 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"root\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-root\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.934671 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-config\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.938949 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-multus-cni-dir\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.942011 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.942194 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-sys\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.942331 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-cnibin\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.942970 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-metrics-client-ca\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.949294 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-proxy\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-main-proxy\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.950566 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/a40bf20d-67a4-45e7-b701-ca5a75376c1f-audit-log\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.950799 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-run-ovn-kubernetes\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.952736 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c0e47249-a5ff-460e-958d-53577387c154-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.953482 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-env-overrides\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.953939 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-system-cni-dir\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.954743 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.955440 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"etc\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-etc\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.956214 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"ovn-ca\" (UniqueName: \"kubernetes.io/configmap/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovn-ca\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.956836 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a40bf20d-67a4-45e7-b701-ca5a75376c1f-tmpfs\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.957040 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-kubelet-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.957101 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-registration-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.957149 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-system-cni-dir\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.957445 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-etc-openvswitch\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.958000 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/05c9a6ea-140a-49a2-bde5-f220b05aa252-config-out\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.958044 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-systemd-units\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.960915 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-cnibin\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.960963 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-node-log\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.960997 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-cni-netd\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.963824 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-config-out\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.964142 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-os-release\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.964775 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-run-openvswitch\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.964826 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-sys\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.965665 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.966200 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/7c63c9be-d1ad-44b6-8b33-09af3ee314af-metrics-client-ca\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.966864 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/05c9a6ea-140a-49a2-bde5-f220b05aa252-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.967596 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.968506 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-wtmp\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.968781 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-run-ovn\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.969229 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"thanos-querier-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c63c9be-d1ad-44b6-8b33-09af3ee314af-thanos-querier-trusted-ca-bundle\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.969773 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"run-systemd-system\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-run-systemd-system\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.970524 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"var-lib-tuned-profiles-data\" (UniqueName: \"kubernetes.io/configmap/91265f01-9dcb-4046-961e-ad4c544224d7-var-lib-tuned-profiles-data\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.970999 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/659b8bb4-81b5-49bf-8127-1d326f30ba2f-host\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.971042 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-run-netns\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.971099 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-os-release\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.972633 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/05c9a6ea-140a-49a2-bde5-f220b05aa252-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.972700 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.973106 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.973322 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3b10ff4-a243-4a4e-b27d-f613e258666d-service-ca-bundle\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.973626 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"plugin-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-plugin-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.977193 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.978157 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovnkube-config\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.978891 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-host\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.979887 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tls\" (UniqueName: \"kubernetes.io/secret/a40bf20d-67a4-45e7-b701-ca5a75376c1f-tls\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.980702 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"prometheus-adapter-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-prometheus-adapter-audit-profiles\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.982371 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-metrics-client-ca\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.982611 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.983336 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/659b8bb4-81b5-49bf-8127-1d326f30ba2f-serviceca\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.994591 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-default-certificate\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.995044 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/05c9a6ea-140a-49a2-bde5-f220b05aa252-tls-assets\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:04.995500 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.006132 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.006470 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c0e47249-a5ff-460e-958d-53577387c154-cni-binary-copy\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.007677 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-serving-certs-ca-bundle\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.010220 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-device-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.011834 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-serving-certs-ca-bundle\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.013413 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-log-socket\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.014766 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.015114 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.017090 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-config-volume\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.017121 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05c9a6ea-140a-49a2-bde5-f220b05aa252-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.017934 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-textfile\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.018622 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"prometheus-adapter-prometheus-config\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-prometheus-adapter-prometheus-config\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.019093 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-rootfs\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.019236 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-slash\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.019443 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8f88cbe5-1221-4e6f-a6c9-f22da152b43f-hosts-file\") pod \"node-resolver-nxzr8\" (UID: \"8f88cbe5-1221-4e6f-a6c9-f22da152b43f\") " pod="openshift-dns/node-resolver-nxzr8" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.019635 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-volume-directive-shadow\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.019856 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-lib-modules\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.020012 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-cni-bin\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.021802 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"telemeter-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-telemeter-trusted-ca-bundle\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.022565 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-var-lib-openvswitch\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.023318 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-openshift-state-metrics-tls\") pod \"openshift-state-metrics-6f5dcdf79b-wq256\" (UID: \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\") " pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.023645 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cookie-secret\" (UniqueName: \"kubernetes.io/secret/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-cookie-secret\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.024279 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-stats-auth\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.024697 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.025104 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.025496 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-state-metrics-tls\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.025798 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-web-config\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.026097 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/88745ab7-efe7-42c0-a2e8-75591518333d-tls-certificates\") pod \"prometheus-operator-admission-webhook-598468bb49-94dsn\" (UID: \"88745ab7-efe7-42c0-a2e8-75591518333d\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.026325 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"non-standard-root-system-trust-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bb84a70-232c-4467-ac0b-647030155908-non-standard-root-system-trust-ca-bundle\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.027243 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-config\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.027602 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-config-volume\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.027744 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-var-run-dbus\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.028448 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-metrics-client-ca\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.028595 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"ovn-cert\" (UniqueName: \"kubernetes.io/secret/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovn-cert\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.029211 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-web-config\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.029562 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-grpc-tls\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.033801 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-kube-etcd-client-certs\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-kube-etcd-client-certs\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.036505 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovn-node-metrics-cert\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.038001 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-metrics-tls\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.039833 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.050588 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfs6g\" (UniqueName: \"kubernetes.io/projected/a40bf20d-67a4-45e7-b701-ca5a75376c1f-kube-api-access-gfs6g\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.050941 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"telemeter-client-tls\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-telemeter-client-tls\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.051194 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.051433 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.051689 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.051858 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.052070 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-proxy-tls\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.052341 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.052568 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-oauth-cookie\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-oauth-cookie\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.061993 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-telemeter-client-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-secret-telemeter-client-kube-rbac-proxy-config\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.071223 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-proxy\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-proxy\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.071510 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-metrics-certs\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.071787 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.072087 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-telemeter-client\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-secret-telemeter-client\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.072530 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.073603 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.074459 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.074696 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-tls\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.078505 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-tls\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.087445 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx2gt\" (UniqueName: \"kubernetes.io/projected/50152296-9489-4fa8-aa42-7858debe1c08-kube-api-access-zx2gt\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.087977 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzzxr\" (UniqueName: \"kubernetes.io/projected/1bb84a70-232c-4467-ac0b-647030155908-kube-api-access-vzzxr\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.088297 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-6f5dcdf79b-wq256\" (UID: \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\") " pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.088599 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfs2p\" (UniqueName: \"kubernetes.io/projected/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-kube-api-access-dfs2p\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.090952 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvz7l\" (UniqueName: \"kubernetes.io/projected/7c63c9be-d1ad-44b6-8b33-09af3ee314af-kube-api-access-pvz7l\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.133518 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7xz4\" (UniqueName: \"kubernetes.io/projected/91265f01-9dcb-4046-961e-ad4c544224d7-kube-api-access-b7xz4\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.134377 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxsm5\" (UniqueName: \"kubernetes.io/projected/05c9a6ea-140a-49a2-bde5-f220b05aa252-kube-api-access-kxsm5\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.141110 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.143744436Z" level=info msg="Running pod sandbox: openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh/POD" id=84bf72ae-8e03-4616-a642-c3d68a94a275 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.144217072Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.146351 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccjjw\" (UniqueName: \"kubernetes.io/projected/c3b10ff4-a243-4a4e-b27d-f613e258666d-kube-api-access-ccjjw\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.152505 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.153138 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.153280960Z" level=info msg="Running pod sandbox: openshift-monitoring/node-exporter-2r6nf/POD" id=f5f20b7d-a2d2-4429-a234-a9b79f7beb0f name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.153387210Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.154489058Z" level=info msg="Running pod sandbox: openshift-ingress/router-default-5cf6bbdff9-tnffs/POD" id=cea2f99d-429b-485e-9f18-1ae520031c8a name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.154699555Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.165871 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmxtc\" (UniqueName: \"kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc\") pod \"network-check-target-xmq2g\" (UID: \"f364a949-09d7-4cfb-83ff-e532b822a557\") " pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.173800 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.174963195Z" level=info msg="Running pod sandbox: openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn/POD" id=b5f01488-c439-4bce-8eba-9abb8505890d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.175098507Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.178939 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.180107088Z" level=info msg="Running pod sandbox: openshift-multus/network-metrics-daemon-hq594/POD" id=3ef0c17d-bba5-488b-ad96-8e18ae386263 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.180153791Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.190849 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.191367 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7www\" (UniqueName: \"kubernetes.io/projected/8f88cbe5-1221-4e6f-a6c9-f22da152b43f-kube-api-access-g7www\") pod \"node-resolver-nxzr8\" (UID: \"8f88cbe5-1221-4e6f-a6c9-f22da152b43f\") " pod="openshift-dns/node-resolver-nxzr8" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.194789 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.199179315Z" level=info msg="Running pod sandbox: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/POD" id=510e0b76-941b-4df9-96cf-95b4608a89ed name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.199229189Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.199776071Z" level=info msg="Running pod sandbox: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/POD" id=7a40a946-0570-4910-85eb-af6d483433f7 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.199826991Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.207061961Z" level=info msg="Got pod network &{Name:prometheus-adapter-5894cc86c7-qhmbh Namespace:openshift-monitoring ID:74f272f4c50be5a640d85c29438b61a9a814b4b072afe86593507911550e6a03 UID:a40bf20d-67a4-45e7-b701-ca5a75376c1f NetNS:/var/run/netns/be37f359-70a2-4ca1-93a2-c054848967c1 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.207213755Z" level=info msg="Adding pod openshift-monitoring_prometheus-adapter-5894cc86c7-qhmbh to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.210098082Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=f5f20b7d-a2d2-4429-a234-a9b79f7beb0f name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:28:05.213303 2927 watcher.go:93] Error while processing event ("/sys/fs/cgroup/devices/kubepods.slice/kubepods-burstable.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/devices/kubepods.slice/kubepods-burstable.slice: no such file or directory Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.216194 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.223519927Z" level=info msg="Running pod sandbox: openshift-monitoring/alertmanager-main-0/POD" id=bc650832-6cc0-4148-bdb8-05b5ae4deaf1 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.223563272Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.226447 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xrks\" (UniqueName: \"kubernetes.io/projected/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-kube-api-access-5xrks\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.230044 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.235527780Z" level=info msg="Running pod sandbox: openshift-cluster-node-tuning-operator/tuned-smwj7/POD" id=1f8ee2d8-864d-438b-bf38-f133bb8f8a04 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.235674743Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.239187 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-98gp9\" (UniqueName: \"kubernetes.io/projected/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-kube-api-access-98gp9\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.241622 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.250837587Z" level=info msg="Got pod network &{Name:prometheus-operator-admission-webhook-598468bb49-94dsn Namespace:openshift-monitoring ID:ff598f9236d4f32150ffef17742e5f55d19d3b2adb4f7800e050a1ea04825617 UID:88745ab7-efe7-42c0-a2e8-75591518333d NetNS:/var/run/netns/dcca133c-c11b-4381-a646-e2e95aca9314 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.250967009Z" level=info msg="Adding pod openshift-monitoring_prometheus-operator-admission-webhook-598468bb49-94dsn to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.263185 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r64p\" (UniqueName: \"kubernetes.io/projected/fc45e847-8f55-4c48-8c48-4bfee2419dca-kube-api-access-4r64p\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.263427636Z" level=info msg="Running pod sandbox: openshift-monitoring/prometheus-k8s-0/POD" id=0e4a725d-2cd9-4f62-8e17-541b1dbb4f78 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.263465012Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:28:05.264153 2927 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa84cbad_3d0f_43bd_a8a9_46dc1cd9c57f.slice/crio-2de3677cadab589863eceefc75ced2bea032cf20d59c2b31ab33b3b6e1993538.scope WatchSource:0}: Error finding container 2de3677cadab589863eceefc75ced2bea032cf20d59c2b31ab33b3b6e1993538: Status 404 returned error can't find the container with id 2de3677cadab589863eceefc75ced2bea032cf20d59c2b31ab33b3b6e1993538 Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.277598977Z" level=info msg="Ran pod sandbox 2de3677cadab589863eceefc75ced2bea032cf20d59c2b31ab33b3b6e1993538 with infra container: openshift-monitoring/node-exporter-2r6nf/POD" id=f5f20b7d-a2d2-4429-a234-a9b79f7beb0f name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.283566523Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180" id=6566fbbe-c118-414d-835e-c3355ffe0fe0 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.283788869Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b4cde205857c1e0a984b38c61275a8370136432c7c84dbfbd7f4835f26bbb935,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180],Size_:297268865,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=6566fbbe-c118-414d-835e-c3355ffe0fe0 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.285028 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-blpwr\" (UniqueName: \"kubernetes.io/projected/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-api-access-blpwr\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.299046486Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180" id=8bb079aa-f06d-4d86-9477-00ea23552834 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.303439 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzfn5\" (UniqueName: \"kubernetes.io/projected/e04a47b4-aa8e-4d4e-9d45-aa8037dcb748-kube-api-access-lzfn5\") pod \"ingress-canary-8xrbh\" (UID: \"e04a47b4-aa8e-4d4e-9d45-aa8037dcb748\") " pod="openshift-ingress-canary/ingress-canary-8xrbh" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.306809484Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b4cde205857c1e0a984b38c61275a8370136432c7c84dbfbd7f4835f26bbb935,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180],Size_:297268865,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=8bb079aa-f06d-4d86-9477-00ea23552834 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.308022167Z" level=info msg="Creating container: openshift-monitoring/node-exporter-2r6nf/init-textfile" id=6058c64f-a889-4588-8166-de82bb12aa40 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.308190627Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.315302440Z" level=info msg="Got pod network &{Name:network-metrics-daemon-hq594 Namespace:openshift-multus ID:1150fb360a7f91f41f9b7eb9840a298cfab29d821843d0d60688bc2927bdfece UID:50152296-9489-4fa8-aa42-7858debe1c08 NetNS:/var/run/netns/e6314a5f-9fa7-4a60-ab76-6473053b4fa4 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.315328815Z" level=info msg="Adding pod openshift-multus_network-metrics-daemon-hq594 to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.329212 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4sjn\" (UniqueName: \"kubernetes.io/projected/659b8bb4-81b5-49bf-8127-1d326f30ba2f-kube-api-access-h4sjn\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.346343860Z" level=info msg="Got pod network &{Name:router-default-5cf6bbdff9-tnffs Namespace:openshift-ingress ID:8a255d5b5b10cf99b23616f05d7fb3c48ec24229e1f6cc5ce6484e928e0a86af UID:c3b10ff4-a243-4a4e-b27d-f613e258666d NetNS:/var/run/netns/e737e306-d06a-4291-bbea-c8e3872fce31 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.346378683Z" level=info msg="Adding pod openshift-ingress_router-default-5cf6bbdff9-tnffs to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.356794 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-g48vj\" (UniqueName: \"kubernetes.io/projected/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-kube-api-access-g48vj\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.358081892Z" level=info msg="Got pod network &{Name:prometheus-k8s-0 Namespace:openshift-monitoring ID:52b90175ef0a6f25e9a947f029d3b67d80d2c31734082f3c71e38fa2ba264d89 UID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 NetNS:/var/run/netns/08082c34-5052-4e18-a998-520794be8216 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.358126372Z" level=info msg="Adding pod openshift-monitoring_prometheus-k8s-0 to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.362500 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-2pfzf" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.374060217Z" level=info msg="Running pod sandbox: openshift-dns/dns-default-2pfzf/POD" id=e1a3ca46-101b-43fa-9721-c16b2c482369 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.374110519Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.377518 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-8xrbh" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.379056 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-wq4k7\" (UniqueName: \"kubernetes.io/projected/c0e47249-a5ff-460e-958d-53577387c154-kube-api-access-wq4k7\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.379281 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.385604654Z" level=info msg="Running pod sandbox: openshift-monitoring/telemeter-client-5598c645c4-mftdv/POD" id=64b71f4f-3284-4db9-a02a-4cd2ff5dd755 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.386389484Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.386173312Z" level=info msg="Running pod sandbox: openshift-ingress-canary/ingress-canary-8xrbh/POD" id=606b9627-db1b-4a3f-874b-b462a6d88dd8 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.387101786Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.388248 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.389671 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-tv6rn\" (UniqueName: \"kubernetes.io/projected/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-kube-api-access-tv6rn\") pod \"multus-pbfzz\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " pod="openshift-multus/multus-pbfzz" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.389892050Z" level=info msg="Running pod sandbox: openshift-multus/multus-additional-cni-plugins-5fgxc/POD" id=0e652b34-5616-44f2-8eb1-15e2937eb87d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.390026151Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.400739816Z" level=info msg="Got pod network &{Name:alertmanager-main-0 Namespace:openshift-monitoring ID:ed0ee704407b6914733819122eb9fc60ca3996f0c1b935170d5903732b869347 UID:05c9a6ea-140a-49a2-bde5-f220b05aa252 NetNS:/var/run/netns/e5957241-af0b-4350-a062-91e5162d6a40 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.400764369Z" level=info msg="Adding pod openshift-monitoring_alertmanager-main-0 to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.404447 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.406709113Z" level=info msg="Running pod sandbox: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/POD" id=0755a791-7501-4417-934f-53470d89bb87 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.406754885Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.406945 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-csks4\" (UniqueName: \"kubernetes.io/projected/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-kube-api-access-csks4\") pod \"ovnkube-node-rp9bd\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.411794 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.413043481Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=1f8ee2d8-864d-438b-bf38-f133bb8f8a04 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.413863124Z" level=info msg="Running pod sandbox: openshift-machine-config-operator/machine-config-daemon-hf8f5/POD" id=57fc77a1-1df5-4f72-bc3e-cf267463f586 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.414001253Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.416522535Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=7a40a946-0570-4910-85eb-af6d483433f7 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.421604 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6g5r\" (UniqueName: \"kubernetes.io/projected/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-kube-api-access-h6g5r\") pod \"openshift-state-metrics-6f5dcdf79b-wq256\" (UID: \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\") " pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.436985 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-nxzr8" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.438070183Z" level=info msg="Running pod sandbox: openshift-dns/node-resolver-nxzr8/POD" id=2b83847c-6fbc-482b-8cfd-a8c4c5a3e740 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.438141357Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.461899641Z" level=info msg="Ran pod sandbox 23c255a61eb7c200b97ccd2916bb330a59475bd2d3e145171f1e9f94a9c27e2e with infra container: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/POD" id=7a40a946-0570-4910-85eb-af6d483433f7 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.464076 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.464411664Z" level=info msg="Ran pod sandbox 0a799b0f1916c3d4fc548b401a6c06c1907860552ea55059042c2484d4fa68dc with infra container: openshift-cluster-node-tuning-operator/tuned-smwj7/POD" id=1f8ee2d8-864d-438b-bf38-f133bb8f8a04 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.470774225Z" level=info msg="Running pod sandbox: openshift-network-diagnostics/network-check-target-xmq2g/POD" id=96524ba2-8daa-4a74-ab1c-222f3b9d9054 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.470810718Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.472043557Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7" id=57e79c84-0650-459d-b42a-15c3d258a584 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.472133524Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ca6ba83315a4882f5c6c4e01c79238e1b794b0e8bfdc2f004990f516e8e0b056" id=c21f0081-c037-4f91-a82f-56a44ca3f6dc name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.477156347Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:6e78a44a973f2298da6f51c1418094403ee8c6c2b57bde9d3285b9e2ac2ba129,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7],Size_:397889589,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=57e79c84-0650-459d-b42a-15c3d258a584 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.477535232Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b2f8382287975a85853941eec2c606723c2ed1c24f59fa46ebdb470b91a481f5,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ca6ba83315a4882f5c6c4e01c79238e1b794b0e8bfdc2f004990f516e8e0b056],Size_:505526574,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=c21f0081-c037-4f91-a82f-56a44ca3f6dc name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.493682427Z" level=info msg="Got pod network &{Name:ingress-canary-8xrbh Namespace:openshift-ingress-canary ID:2c2a604331152b239cd95403d2bb266e2ca816b6a0414973001dc448b2d2b416 UID:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748 NetNS:/var/run/netns/28d99893-e7fa-4b86-9cf0-5e450f5457d2 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.493823342Z" level=info msg="Adding pod openshift-ingress-canary_ingress-canary-8xrbh to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.507688 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-xw6d2" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.511802799Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7" id=e544539f-bc19-4a17-80b7-a1feb1eabdf4 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.512086932Z" level=info msg="Running pod sandbox: openshift-image-registry/node-ca-xw6d2/POD" id=72b2549b-2e41-483b-9bf6-7b848149d2e7 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.512128218Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.526971779Z" level=info msg="Got pod network &{Name:telemeter-client-5598c645c4-mftdv Namespace:openshift-monitoring ID:a16c8f8a13d2354862769ea877c06e032be034fdb676b065e5595fbceae680cf UID:fc45e847-8f55-4c48-8c48-4bfee2419dca NetNS:/var/run/netns/96790582-8aca-42b0-a796-fa0aa2193bfa Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.526997372Z" level=info msg="Adding pod openshift-monitoring_telemeter-client-5598c645c4-mftdv to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.527358630Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ca6ba83315a4882f5c6c4e01c79238e1b794b0e8bfdc2f004990f516e8e0b056" id=2134a685-d439-4661-aa66-257c3a105bd2 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.528518863Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b2f8382287975a85853941eec2c606723c2ed1c24f59fa46ebdb470b91a481f5,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ca6ba83315a4882f5c6c4e01c79238e1b794b0e8bfdc2f004990f516e8e0b056],Size_:505526574,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=2134a685-d439-4661-aa66-257c3a105bd2 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.531056180Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:6e78a44a973f2298da6f51c1418094403ee8c6c2b57bde9d3285b9e2ac2ba129,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7],Size_:397889589,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=e544539f-bc19-4a17-80b7-a1feb1eabdf4 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.532051195Z" level=info msg="Creating container: openshift-cluster-node-tuning-operator/tuned-smwj7/tuned" id=540673ed-b22d-4376-adbc-d0cdfafe4f6a name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.532292822Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.547732980Z" level=info msg="Creating container: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-driver" id=f35c7b26-edff-48ac-baed-49ceb42ef2d5 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.547972756Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.548552698Z" level=info msg="Got pod network &{Name:dns-default-2pfzf Namespace:openshift-dns ID:dcf02c2c8b254c0ba075cbb4447c97111fedbd18176c9b10dcf0b8e0fa4c2b1d UID:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 NetNS:/var/run/netns/f79bd825-de4b-4a51-9fe9-dc217b8ab2c4 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.551762427Z" level=info msg="Adding pod openshift-dns_dns-default-2pfzf to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.563544 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-pbfzz" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.567560174Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=0e652b34-5616-44f2-8eb1-15e2937eb87d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.583373502Z" level=info msg="Running pod sandbox: openshift-multus/multus-pbfzz/POD" id=5c296666-996d-4f19-b278-1ce583629371 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.585164474Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.588014 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.592620726Z" level=info msg="Running pod sandbox: openshift-ovn-kubernetes/ovnkube-node-rp9bd/POD" id=62774430-cfeb-439c-ba8e-f49c12764335 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.593594013Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.626891100Z" level=info msg="Created container 8e22fac0016ec250c09e2029cb09e2f40137c05c12f99ffc18c099137ec5745d: openshift-monitoring/node-exporter-2r6nf/init-textfile" id=6058c64f-a889-4588-8166-de82bb12aa40 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:28:05.627408 2927 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0e47249_a5ff_460e_958d_53577387c154.slice/crio-8ea5e7e9fb15db146766a6199273736a81c0e10f5714578c6922f99837137af4.scope WatchSource:0}: Error finding container 8ea5e7e9fb15db146766a6199273736a81c0e10f5714578c6922f99837137af4: Status 404 returned error can't find the container with id 8ea5e7e9fb15db146766a6199273736a81c0e10f5714578c6922f99837137af4 Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.627847335Z" level=info msg="Starting container: 8e22fac0016ec250c09e2029cb09e2f40137c05c12f99ffc18c099137ec5745d" id=1258f17b-65d4-4547-b65e-d7f9ec5e651b name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.632025703Z" level=info msg="Ran pod sandbox 8ea5e7e9fb15db146766a6199273736a81c0e10f5714578c6922f99837137af4 with infra container: openshift-multus/multus-additional-cni-plugins-5fgxc/POD" id=0e652b34-5616-44f2-8eb1-15e2937eb87d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.632985641Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:3fc008580e4e404960b2de4020a2939ab6421bdf8f51c0b60af4ad903272c741" id=87133437-b8ae-42cf-ab00-adf1636867ef name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.634019675Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:6bb2a44a1f0506c18ef471cac7f7e8991d1e6b9d649ccbce50f751dd2863d06d,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:3fc008580e4e404960b2de4020a2939ab6421bdf8f51c0b60af4ad903272c741],Size_:382411122,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=87133437-b8ae-42cf-ab00-adf1636867ef name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.636419055Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:3fc008580e4e404960b2de4020a2939ab6421bdf8f51c0b60af4ad903272c741" id=b9919886-24c9-42d8-9f54-63e1580b55ef name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.636574922Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:6bb2a44a1f0506c18ef471cac7f7e8991d1e6b9d649ccbce50f751dd2863d06d,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:3fc008580e4e404960b2de4020a2939ab6421bdf8f51c0b60af4ad903272c741],Size_:382411122,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b9919886-24c9-42d8-9f54-63e1580b55ef name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.637139617Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/egress-router-binary-copy" id=eb2eb774-e9ec-42a8-a53b-10209d83d3df name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.637238151Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.644900856Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=2b83847c-6fbc-482b-8cfd-a8c4c5a3e740 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.647893721Z" level=info msg="Started container" PID=3207 containerID=8e22fac0016ec250c09e2029cb09e2f40137c05c12f99ffc18c099137ec5745d description=openshift-monitoring/node-exporter-2r6nf/init-textfile id=1258f17b-65d4-4547-b65e-d7f9ec5e651b name=/runtime.v1.RuntimeService/StartContainer sandboxID=2de3677cadab589863eceefc75ced2bea032cf20d59c2b31ab33b3b6e1993538 Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.661845573Z" level=info msg="Got pod network &{Name:kube-state-metrics-6798b548b9-xl88q Namespace:openshift-monitoring ID:a937bd9597c90954ad3041a6e3670bd2dfc453dbbd7b4d1b6d48e35dae7b83c0 UID:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e NetNS:/var/run/netns/7dcc1407-51e3-46a5-8bb4-1903ac68eef7 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.661877020Z" level=info msg="Adding pod openshift-monitoring_kube-state-metrics-6798b548b9-xl88q to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.665348893Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=57fc77a1-1df5-4f72-bc3e-cf267463f586 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.667048825Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=72b2549b-2e41-483b-9bf6-7b848149d2e7 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.687522420Z" level=info msg="Got pod network &{Name:network-check-target-xmq2g Namespace:openshift-network-diagnostics ID:68c40ad6338d492b4aa5a2126d3afd81d13c48ead9d59f5b59bd6ce18b97e366 UID:f364a949-09d7-4cfb-83ff-e532b822a557 NetNS:/var/run/netns/eee4c6a8-243a-4fce-8d4b-4475692c8e09 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.687712419Z" level=info msg="Adding pod openshift-network-diagnostics_network-check-target-xmq2g to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:28:05.697470 2927 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f88cbe5_1221_4e6f_a6c9_f22da152b43f.slice/crio-a22e1d52a296a877749d67bb8347e63d619ee2475777ed73628f6d33736e68d4.scope WatchSource:0}: Error finding container a22e1d52a296a877749d67bb8347e63d619ee2475777ed73628f6d33736e68d4: Status 404 returned error can't find the container with id a22e1d52a296a877749d67bb8347e63d619ee2475777ed73628f6d33736e68d4 Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.710418710Z" level=info msg="Ran pod sandbox a22e1d52a296a877749d67bb8347e63d619ee2475777ed73628f6d33736e68d4 with infra container: openshift-dns/node-resolver-nxzr8/POD" id=2b83847c-6fbc-482b-8cfd-a8c4c5a3e740 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.717053803Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f93a29f8d7455a5067ecde85394835bda877174193ab693b2b010d96c8cd9183" id=6fdcb2f2-42bc-42d1-9c27-3ee116e1a1bc name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.717394580Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:9b1e063a0faa2683ff00ee1f8b324af985c1c142530a63c7f8729ec8dab01b38,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f93a29f8d7455a5067ecde85394835bda877174193ab693b2b010d96c8cd9183],Size_:480731528,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=6fdcb2f2-42bc-42d1-9c27-3ee116e1a1bc name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.728107 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.733084879Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f93a29f8d7455a5067ecde85394835bda877174193ab693b2b010d96c8cd9183" id=987c05fa-95d0-4905-941c-1f9e7b3eca7a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:28:05.733265 2927 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod659b8bb4_81b5_49bf_8127_1d326f30ba2f.slice/crio-f16d1d77b76cea9d449631cf97d4fe890b2fee1b5dc69a019689afaeec8f7b05.scope WatchSource:0}: Error finding container f16d1d77b76cea9d449631cf97d4fe890b2fee1b5dc69a019689afaeec8f7b05: Status 404 returned error can't find the container with id f16d1d77b76cea9d449631cf97d4fe890b2fee1b5dc69a019689afaeec8f7b05 Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.733516795Z" level=info msg="Running pod sandbox: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/POD" id=ff04d155-666b-4a8f-ba4b-1d1da011fa66 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.733570637Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:28:05.735509 2927 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod997b7ab4_939e_465c_9c7d_4a2ebe3a797a.slice/crio-c42a602a42d67c9af5ab623c3df747e5144697eb541c212fccf9ff7c0f6542d7.scope WatchSource:0}: Error finding container c42a602a42d67c9af5ab623c3df747e5144697eb541c212fccf9ff7c0f6542d7: Status 404 returned error can't find the container with id c42a602a42d67c9af5ab623c3df747e5144697eb541c212fccf9ff7c0f6542d7 Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.754763643Z" level=info msg="Ran pod sandbox f16d1d77b76cea9d449631cf97d4fe890b2fee1b5dc69a019689afaeec8f7b05 with infra container: openshift-image-registry/node-ca-xw6d2/POD" id=72b2549b-2e41-483b-9bf6-7b848149d2e7 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.757172121Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:9b1e063a0faa2683ff00ee1f8b324af985c1c142530a63c7f8729ec8dab01b38,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f93a29f8d7455a5067ecde85394835bda877174193ab693b2b010d96c8cd9183],Size_:480731528,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=987c05fa-95d0-4905-941c-1f9e7b3eca7a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.759894612Z" level=info msg="Ran pod sandbox c42a602a42d67c9af5ab623c3df747e5144697eb541c212fccf9ff7c0f6542d7 with infra container: openshift-machine-config-operator/machine-config-daemon-hf8f5/POD" id=57fc77a1-1df5-4f72-bc3e-cf267463f586 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.764477626Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8" id=f45fb106-9381-48d1-8c59-be2c478dcd6f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.764780780Z" level=info msg="Creating container: openshift-dns/node-resolver-nxzr8/dns-node-resolver" id=09561107-8ef0-40c3-a7da-00473ed7519f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.765156445Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.773485871Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cb9edf2251410c340a6125ff331924ab57a72ad3166bd9e1305b6d3030284990" id=d4ec050c-4fb4-45dc-bac6-688cc3b9a02c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.776162212Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2b208acf1f002c143fdb248dbee2517f7b6d168def4cfef1f3188939123d0dc3,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cb9edf2251410c340a6125ff331924ab57a72ad3166bd9e1305b6d3030284990],Size_:536350945,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=d4ec050c-4fb4-45dc-bac6-688cc3b9a02c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.775367280Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:c9e2954d8b9d9e1594d0a16b90086a9297ffd65773a6aa4a64f325e871ce83d1,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8],Size_:363009057,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=f45fb106-9381-48d1-8c59-be2c478dcd6f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.778136388Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cb9edf2251410c340a6125ff331924ab57a72ad3166bd9e1305b6d3030284990" id=87b5bb59-c59f-4001-90cc-8cb1db216747 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.778567933Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2b208acf1f002c143fdb248dbee2517f7b6d168def4cfef1f3188939123d0dc3,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cb9edf2251410c340a6125ff331924ab57a72ad3166bd9e1305b6d3030284990],Size_:536350945,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=87b5bb59-c59f-4001-90cc-8cb1db216747 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.780316018Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8" id=c57691bf-1565-40a9-b10e-3cb57db2ddbf name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.780524253Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:c9e2954d8b9d9e1594d0a16b90086a9297ffd65773a6aa4a64f325e871ce83d1,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8],Size_:363009057,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=c57691bf-1565-40a9-b10e-3cb57db2ddbf name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.784892880Z" level=info msg="Creating container: openshift-machine-config-operator/machine-config-daemon-hf8f5/machine-config-daemon" id=9cd843e9-cfc2-47c6-90c9-14bac58e251f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.785151426Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.785624334Z" level=info msg="Creating container: openshift-image-registry/node-ca-xw6d2/node-ca" id=7c981693-bd77-48db-888f-52d9417ee4aa name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.785764743Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.848436342Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=5c296666-996d-4f19-b278-1ce583629371 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.901534946Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=62774430-cfeb-439c-ba8e-f49c12764335 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.902210879Z" level=info msg="Created container eb2903e092d65e9ea578f6d3048aa2bcf5b3ed144f88c5b61a8a443631ae7f6b: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-driver" id=f35c7b26-edff-48ac-baed-49ceb42ef2d5 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.902753215Z" level=info msg="Starting container: eb2903e092d65e9ea578f6d3048aa2bcf5b3ed144f88c5b61a8a443631ae7f6b" id=61fe5d4c-5f12-4794-994d-969cee88f9d2 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:28:05.914925 2927 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf907add9_a2cf_4fc7_b7f8_f283bc1c2dd7.slice/crio-bcac6a2d9a941220c56968841333b62aa70ad0818ad3c6858d36fafbd613d3b1.scope WatchSource:0}: Error finding container bcac6a2d9a941220c56968841333b62aa70ad0818ad3c6858d36fafbd613d3b1: Status 404 returned error can't find the container with id bcac6a2d9a941220c56968841333b62aa70ad0818ad3c6858d36fafbd613d3b1 Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.918448970Z" level=info msg="Ran pod sandbox bcac6a2d9a941220c56968841333b62aa70ad0818ad3c6858d36fafbd613d3b1 with infra container: openshift-multus/multus-pbfzz/POD" id=5c296666-996d-4f19-b278-1ce583629371 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.921182882Z" level=info msg="Created container 66a74c20cf61babc17acec6afbc9d04f43a3d23e3748c968ae7871b360e4b2f6: openshift-cluster-node-tuning-operator/tuned-smwj7/tuned" id=540673ed-b22d-4376-adbc-d0cdfafe4f6a name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:28:05.926339 2927 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaac1c4d4_278d_4a53_b2f7_a75ba43c2d18.slice/crio-fd9deb131d1f752a07a6704193bcd1cf1dab19ff2e6439be2d911f94c338da51.scope WatchSource:0}: Error finding container fd9deb131d1f752a07a6704193bcd1cf1dab19ff2e6439be2d911f94c338da51: Status 404 returned error can't find the container with id fd9deb131d1f752a07a6704193bcd1cf1dab19ff2e6439be2d911f94c338da51 Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.929501051Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=fed1042e-82ad-4f32-ae98-b5dc06764084 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.929579163Z" level=info msg="Starting container: 66a74c20cf61babc17acec6afbc9d04f43a3d23e3748c968ae7871b360e4b2f6" id=23125028-c4d7-4cb0-a1e8-f8088f95e46e name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.954491481Z" level=info msg="Ran pod sandbox fd9deb131d1f752a07a6704193bcd1cf1dab19ff2e6439be2d911f94c338da51 with infra container: openshift-ovn-kubernetes/ovnkube-node-rp9bd/POD" id=62774430-cfeb-439c-ba8e-f49c12764335 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.955464639Z" level=info msg="Started container" PID=3309 containerID=eb2903e092d65e9ea578f6d3048aa2bcf5b3ed144f88c5b61a8a443631ae7f6b description=openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-driver id=61fe5d4c-5f12-4794-994d-969cee88f9d2 name=/runtime.v1.RuntimeService/StartContainer sandboxID=23c255a61eb7c200b97ccd2916bb330a59475bd2d3e145171f1e9f94a9c27e2e Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.969800935Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=fed1042e-82ad-4f32-ae98-b5dc06764084 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.976180013Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:96010fa3364bd5030d4445bbc80df5c8ec18d2b5579f23f71235bcda47b38ca6" id=dea62dca-b65d-4602-a821-af7cedb18b6a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.976833768Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:96010fa3364bd5030d4445bbc80df5c8ec18d2b5579f23f71235bcda47b38ca6 not found" id=dea62dca-b65d-4602-a821-af7cedb18b6a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.978465160Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=cad91a5e-15f0-450a-804d-d6ad9d95ef23 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:05.980209 2927 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.980160948Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=cad91a5e-15f0-450a-804d-d6ad9d95ef23 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:05.985837790Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:96010fa3364bd5030d4445bbc80df5c8ec18d2b5579f23f71235bcda47b38ca6" id=c719d34b-89a5-4d84-adde-2dcf5972639b name=/runtime.v1.ImageService/PullImage Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.002496176Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:96010fa3364bd5030d4445bbc80df5c8ec18d2b5579f23f71235bcda47b38ca6\"" Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.008964155Z" level=info msg="Started container" PID=3308 containerID=66a74c20cf61babc17acec6afbc9d04f43a3d23e3748c968ae7871b360e4b2f6 description=openshift-cluster-node-tuning-operator/tuned-smwj7/tuned id=23125028-c4d7-4cb0-a1e8-f8088f95e46e name=/runtime.v1.RuntimeService/StartContainer sandboxID=0a799b0f1916c3d4fc548b401a6c06c1907860552ea55059042c2484d4fa68dc Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.013078283Z" level=info msg="Creating container: openshift-multus/multus-pbfzz/kube-multus" id=eb429a13-f132-4675-8578-7d75b200559f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.013828538Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.231422160Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:72c764b55efc147864960930ae05b7928c5f976eb76820b38997f0df7afab820" id=10571487-ba4e-4068-885b-83d434845c04 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.231691946Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:0bb939dc4e915f972f52c355cea411db098a9c605b2ca4bc1048447ae297391e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:72c764b55efc147864960930ae05b7928c5f976eb76820b38997f0df7afab820],Size_:293414029,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=10571487-ba4e-4068-885b-83d434845c04 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.244172823Z" level=info msg="Got pod network &{Name:openshift-state-metrics-6f5dcdf79b-wq256 Namespace:openshift-monitoring ID:afb580e701142b75f062c265b00883e2d5567b37000d823ade7ce8eb8a057b02 UID:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 NetNS:/var/run/netns/66f76849-1ec6-4745-9cd9-8c97630f69f8 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.244350501Z" level=info msg="Adding pod openshift-monitoring_openshift-state-metrics-6f5dcdf79b-wq256 to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.308856263Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:72c764b55efc147864960930ae05b7928c5f976eb76820b38997f0df7afab820" id=731d9641-78e3-49d7-bece-1c22cf8c7436 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.310415552Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:0bb939dc4e915f972f52c355cea411db098a9c605b2ca4bc1048447ae297391e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:72c764b55efc147864960930ae05b7928c5f976eb76820b38997f0df7afab820],Size_:293414029,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=731d9641-78e3-49d7-bece-1c22cf8c7436 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.318130172Z" level=info msg="Creating container: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-node-driver-registrar" id=e0c7b8b1-7020-47c0-982b-602cbadbcd10 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.318445598Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:06.324964 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" event=&{ID:997b7ab4-939e-465c-9c7d-4a2ebe3a797a Type:ContainerStarted Data:c42a602a42d67c9af5ab623c3df747e5144697eb541c212fccf9ff7c0f6542d7} Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:06.341017 2927 generic.go:296] "Generic (PLEG): container finished" podID=aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f containerID="8e22fac0016ec250c09e2029cb09e2f40137c05c12f99ffc18c099137ec5745d" exitCode=0 Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:06.341090 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-2r6nf" event=&{ID:aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f Type:ContainerDied Data:8e22fac0016ec250c09e2029cb09e2f40137c05c12f99ffc18c099137ec5745d} Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:06.341125 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-2r6nf" event=&{ID:aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f Type:ContainerStarted Data:2de3677cadab589863eceefc75ced2bea032cf20d59c2b31ab33b3b6e1993538} Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.342948923Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180" id=0fb74baa-ec86-430e-93ee-b9792d8b3b2e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:06.354875 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" event=&{ID:aac1c4d4-278d-4a53-b2f7-a75ba43c2d18 Type:ContainerStarted Data:fd9deb131d1f752a07a6704193bcd1cf1dab19ff2e6439be2d911f94c338da51} Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:06.356294 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-xw6d2" event=&{ID:659b8bb4-81b5-49bf-8127-1d326f30ba2f Type:ContainerStarted Data:f16d1d77b76cea9d449631cf97d4fe890b2fee1b5dc69a019689afaeec8f7b05} Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.357892502Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b4cde205857c1e0a984b38c61275a8370136432c7c84dbfbd7f4835f26bbb935,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180],Size_:297268865,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=0fb74baa-ec86-430e-93ee-b9792d8b3b2e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.360142200Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180" id=27e0e6d2-2f45-435c-9b25-04d188f65eaf name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.366166747Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b4cde205857c1e0a984b38c61275a8370136432c7c84dbfbd7f4835f26bbb935,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180],Size_:297268865,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=27e0e6d2-2f45-435c-9b25-04d188f65eaf name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.378300784Z" level=info msg="Creating container: openshift-monitoring/node-exporter-2r6nf/node-exporter" id=11ae6bd5-927f-486c-bb19-1eb7a01c83bb name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.378583955Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:06.380184 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" event=&{ID:1bb84a70-232c-4467-ac0b-647030155908 Type:ContainerStarted Data:eb2903e092d65e9ea578f6d3048aa2bcf5b3ed144f88c5b61a8a443631ae7f6b} Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:06.380221 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" event=&{ID:1bb84a70-232c-4467-ac0b-647030155908 Type:ContainerStarted Data:23c255a61eb7c200b97ccd2916bb330a59475bd2d3e145171f1e9f94a9c27e2e} Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:06.381277 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerStarted Data:8ea5e7e9fb15db146766a6199273736a81c0e10f5714578c6922f99837137af4} Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:06.384852 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-smwj7" event=&{ID:91265f01-9dcb-4046-961e-ad4c544224d7 Type:ContainerStarted Data:66a74c20cf61babc17acec6afbc9d04f43a3d23e3748c968ae7871b360e4b2f6} Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:06.384887 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-smwj7" event=&{ID:91265f01-9dcb-4046-961e-ad4c544224d7 Type:ContainerStarted Data:0a799b0f1916c3d4fc548b401a6c06c1907860552ea55059042c2484d4fa68dc} Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:06.387102 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-nxzr8" event=&{ID:8f88cbe5-1221-4e6f-a6c9-f22da152b43f Type:ContainerStarted Data:a22e1d52a296a877749d67bb8347e63d619ee2475777ed73628f6d33736e68d4} Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:06.388564 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-pbfzz" event=&{ID:f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7 Type:ContainerStarted Data:bcac6a2d9a941220c56968841333b62aa70ad0818ad3c6858d36fafbd613d3b1} Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.738346272Z" level=info msg="Created container b152d575dba3f481cd90bee054d196f878881ffa033e5b93e2cc659f5eb546a0: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-node-driver-registrar" id=e0c7b8b1-7020-47c0-982b-602cbadbcd10 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.772057138Z" level=info msg="Starting container: b152d575dba3f481cd90bee054d196f878881ffa033e5b93e2cc659f5eb546a0" id=3202382c-743d-4436-9e69-06ce826f14ba name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.784006107Z" level=info msg="Created container f732c4ec2507c71796c1af8b07ed9b67aa246c5803e0c0b5a9081c58c6ae0831: openshift-monitoring/node-exporter-2r6nf/node-exporter" id=11ae6bd5-927f-486c-bb19-1eb7a01c83bb name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.785960488Z" level=info msg="Starting container: f732c4ec2507c71796c1af8b07ed9b67aa246c5803e0c0b5a9081c58c6ae0831" id=ade9d603-d6d7-4e1d-9ee6-35f6bdec7669 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.787371788Z" level=info msg="Created container 34bd660952ff411e0d0fba5ac3a51504a5295d0629a458a9b1358a8fce37d46b: openshift-machine-config-operator/machine-config-daemon-hf8f5/machine-config-daemon" id=9cd843e9-cfc2-47c6-90c9-14bac58e251f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.787781710Z" level=info msg="Starting container: 34bd660952ff411e0d0fba5ac3a51504a5295d0629a458a9b1358a8fce37d46b" id=11a83e63-fcc8-4972-ab93-cc249195dbc0 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.791971730Z" level=info msg="Started container" PID=3466 containerID=b152d575dba3f481cd90bee054d196f878881ffa033e5b93e2cc659f5eb546a0 description=openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-node-driver-registrar id=3202382c-743d-4436-9e69-06ce826f14ba name=/runtime.v1.RuntimeService/StartContainer sandboxID=23c255a61eb7c200b97ccd2916bb330a59475bd2d3e145171f1e9f94a9c27e2e Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.815594670Z" level=info msg="Started container" PID=3469 containerID=f732c4ec2507c71796c1af8b07ed9b67aa246c5803e0c0b5a9081c58c6ae0831 description=openshift-monitoring/node-exporter-2r6nf/node-exporter id=ade9d603-d6d7-4e1d-9ee6-35f6bdec7669 name=/runtime.v1.RuntimeService/StartContainer sandboxID=2de3677cadab589863eceefc75ced2bea032cf20d59c2b31ab33b3b6e1993538 Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.875376126Z" level=info msg="Started container" PID=3479 containerID=34bd660952ff411e0d0fba5ac3a51504a5295d0629a458a9b1358a8fce37d46b description=openshift-machine-config-operator/machine-config-daemon-hf8f5/machine-config-daemon id=11a83e63-fcc8-4972-ab93-cc249195dbc0 name=/runtime.v1.RuntimeService/StartContainer sandboxID=c42a602a42d67c9af5ab623c3df747e5144697eb541c212fccf9ff7c0f6542d7 Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.882046458Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ea25c2f3749dee6981359b4655a9acfcd4d95547066ccdf328354b548da74932" id=7f5ec1dd-3da5-4fd2-9378-1c4e4d2371d1 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.882393378Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:58a5591f9334b70902e27f9a172f3e141b8f07a879507b850a76c2499014eec9,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ea25c2f3749dee6981359b4655a9acfcd4d95547066ccdf328354b548da74932],Size_:291396124,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=7f5ec1dd-3da5-4fd2-9378-1c4e4d2371d1 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.886472650Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=48b3f51c-2869-46e5-a38b-724a75e0c8ed name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.886732040Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=48b3f51c-2869-46e5-a38b-724a75e0c8ed name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.886895553Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ea25c2f3749dee6981359b4655a9acfcd4d95547066ccdf328354b548da74932" id=7c5ed7f9-7557-4192-9c10-094f4d3eb88c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.887113764Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:58a5591f9334b70902e27f9a172f3e141b8f07a879507b850a76c2499014eec9,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ea25c2f3749dee6981359b4655a9acfcd4d95547066ccdf328354b548da74932],Size_:291396124,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=7c5ed7f9-7557-4192-9c10-094f4d3eb88c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.888633214Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=2023cbdf-f4a8-4f02-9dfa-1380c432f752 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.888821223Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=2023cbdf-f4a8-4f02-9dfa-1380c432f752 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.890489745Z" level=info msg="Creating container: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-liveness-probe" id=d0088c4e-f165-4f59-98de-9ff59a1e38eb name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.890823293Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.891111703Z" level=info msg="Creating container: openshift-monitoring/node-exporter-2r6nf/kube-rbac-proxy" id=1980cbd9-f98b-4695-b752-8a68246adaa4 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.891208886Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.943745557Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=bbd4cdb6-6536-4946-9cd9-0bf81a16c561 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.946528939Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=bbd4cdb6-6536-4946-9cd9-0bf81a16c561 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.955897834Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=c7d14ef8-fcf6-44dd-876b-f9c13a615ea8 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.956390745Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=c7d14ef8-fcf6-44dd-876b-f9c13a615ea8 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.969227853Z" level=info msg="Creating container: openshift-machine-config-operator/machine-config-daemon-hf8f5/oauth-proxy" id=2858a90e-d121-492b-aa0b-1310030dd5be name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:06.969449791Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:07.228099268Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:96010fa3364bd5030d4445bbc80df5c8ec18d2b5579f23f71235bcda47b38ca6\"" Jan 05 09:28:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:07.282177047Z" level=info msg="Created container ed90e7004256ef31252e48e33ab49e052cce0d135c4b39e424df92d25282d181: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-liveness-probe" id=d0088c4e-f165-4f59-98de-9ff59a1e38eb name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:07.283501320Z" level=info msg="Starting container: ed90e7004256ef31252e48e33ab49e052cce0d135c4b39e424df92d25282d181" id=92b1c129-e0ac-49b4-bde9-fdad5d62a6ce name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:28:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:07.310762394Z" level=info msg="Started container" PID=3595 containerID=ed90e7004256ef31252e48e33ab49e052cce0d135c4b39e424df92d25282d181 description=openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-liveness-probe id=92b1c129-e0ac-49b4-bde9-fdad5d62a6ce name=/runtime.v1.RuntimeService/StartContainer sandboxID=23c255a61eb7c200b97ccd2916bb330a59475bd2d3e145171f1e9f94a9c27e2e Jan 05 09:28:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:07.393105 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" event=&{ID:997b7ab4-939e-465c-9c7d-4a2ebe3a797a Type:ContainerStarted Data:34bd660952ff411e0d0fba5ac3a51504a5295d0629a458a9b1358a8fce37d46b} Jan 05 09:28:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:07.397288 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-2r6nf" event=&{ID:aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f Type:ContainerStarted Data:f732c4ec2507c71796c1af8b07ed9b67aa246c5803e0c0b5a9081c58c6ae0831} Jan 05 09:28:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:07.403963 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" event=&{ID:1bb84a70-232c-4467-ac0b-647030155908 Type:ContainerStarted Data:ed90e7004256ef31252e48e33ab49e052cce0d135c4b39e424df92d25282d181} Jan 05 09:28:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:07.404306 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" event=&{ID:1bb84a70-232c-4467-ac0b-647030155908 Type:ContainerStarted Data:b152d575dba3f481cd90bee054d196f878881ffa033e5b93e2cc659f5eb546a0} Jan 05 09:28:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:07.780876516Z" level=info msg="Created container e40d040cc9cf721f176d342346bcd293f08fab396828b9dd6027d9f858fb489b: openshift-image-registry/node-ca-xw6d2/node-ca" id=7c981693-bd77-48db-888f-52d9417ee4aa name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:07.782083276Z" level=info msg="Created container 0dde32b8bba42d827b7f9b8e7f3f00a32511565f8a3b5fbdb7328018feba4c3a: openshift-machine-config-operator/machine-config-daemon-hf8f5/oauth-proxy" id=2858a90e-d121-492b-aa0b-1310030dd5be name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:07.793532689Z" level=info msg="Starting container: 0dde32b8bba42d827b7f9b8e7f3f00a32511565f8a3b5fbdb7328018feba4c3a" id=cb1a21b9-4aec-4852-80fe-745f78991ffe name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:28:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:07.798990624Z" level=info msg="Starting container: e40d040cc9cf721f176d342346bcd293f08fab396828b9dd6027d9f858fb489b" id=4156bdd5-0da8-4d03-bcc3-2377e9d3060b name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:28:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:07.894852848Z" level=info msg="Created container bcf5ae9d6717d0c7922cf350cf9cecc4e5443de30e41d454fff49d818f809d14: openshift-monitoring/node-exporter-2r6nf/kube-rbac-proxy" id=1980cbd9-f98b-4695-b752-8a68246adaa4 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:07.899563884Z" level=info msg="Starting container: bcf5ae9d6717d0c7922cf350cf9cecc4e5443de30e41d454fff49d818f809d14" id=a6726fe6-f7d1-404a-8290-0da81df507f8 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:28:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:07.914100247Z" level=info msg="Started container" PID=3644 containerID=0dde32b8bba42d827b7f9b8e7f3f00a32511565f8a3b5fbdb7328018feba4c3a description=openshift-machine-config-operator/machine-config-daemon-hf8f5/oauth-proxy id=cb1a21b9-4aec-4852-80fe-745f78991ffe name=/runtime.v1.RuntimeService/StartContainer sandboxID=c42a602a42d67c9af5ab623c3df747e5144697eb541c212fccf9ff7c0f6542d7 Jan 05 09:28:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:07.947708695Z" level=info msg="Started container" PID=3620 containerID=bcf5ae9d6717d0c7922cf350cf9cecc4e5443de30e41d454fff49d818f809d14 description=openshift-monitoring/node-exporter-2r6nf/kube-rbac-proxy id=a6726fe6-f7d1-404a-8290-0da81df507f8 name=/runtime.v1.RuntimeService/StartContainer sandboxID=2de3677cadab589863eceefc75ced2bea032cf20d59c2b31ab33b3b6e1993538 Jan 05 09:28:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:07.976032136Z" level=info msg="Started container" PID=3651 containerID=e40d040cc9cf721f176d342346bcd293f08fab396828b9dd6027d9f858fb489b description=openshift-image-registry/node-ca-xw6d2/node-ca id=4156bdd5-0da8-4d03-bcc3-2377e9d3060b name=/runtime.v1.RuntimeService/StartContainer sandboxID=f16d1d77b76cea9d449631cf97d4fe890b2fee1b5dc69a019689afaeec8f7b05 Jan 05 09:28:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:08.066552737Z" level=info msg="Created container da19cbb49f674958c161a6b6f29875eb5d015a5a48e2c25147c41964dd2f2f0c: openshift-multus/multus-pbfzz/kube-multus" id=eb429a13-f132-4675-8578-7d75b200559f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:08.069042400Z" level=info msg="Starting container: da19cbb49f674958c161a6b6f29875eb5d015a5a48e2c25147c41964dd2f2f0c" id=412f3971-5e4e-4460-a471-be59957a8371 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:28:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:08.102609731Z" level=info msg="Created container 24638f270f12e2b431e69ccfc44265012351e11345c2e8b283d850d777824b59: openshift-dns/node-resolver-nxzr8/dns-node-resolver" id=09561107-8ef0-40c3-a7da-00473ed7519f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:08.119463696Z" level=info msg="Starting container: 24638f270f12e2b431e69ccfc44265012351e11345c2e8b283d850d777824b59" id=95ca71a8-450f-4143-b35b-bbd11b2f375e name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:28:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:08.138091002Z" level=info msg="Started container" PID=3727 containerID=da19cbb49f674958c161a6b6f29875eb5d015a5a48e2c25147c41964dd2f2f0c description=openshift-multus/multus-pbfzz/kube-multus id=412f3971-5e4e-4460-a471-be59957a8371 name=/runtime.v1.RuntimeService/StartContainer sandboxID=bcac6a2d9a941220c56968841333b62aa70ad0818ad3c6858d36fafbd613d3b1 Jan 05 09:28:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:08.166884780Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_dc316480-6f54-41af-9926-89e3414d4400\"" Jan 05 09:28:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:08.171236470Z" level=info msg="Created container 981d507021d63e67275789bb4c4c1625210c441f07a1e6819d475fb1eb066879: openshift-multus/multus-additional-cni-plugins-5fgxc/egress-router-binary-copy" id=eb2eb774-e9ec-42a8-a53b-10209d83d3df name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:08.179956801Z" level=info msg="Starting container: 981d507021d63e67275789bb4c4c1625210c441f07a1e6819d475fb1eb066879" id=9ec47448-93f9-421a-a61e-072ddaec1ebc name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:28:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:08.184016628Z" level=info msg="Started container" PID=3724 containerID=24638f270f12e2b431e69ccfc44265012351e11345c2e8b283d850d777824b59 description=openshift-dns/node-resolver-nxzr8/dns-node-resolver id=95ca71a8-450f-4143-b35b-bbd11b2f375e name=/runtime.v1.RuntimeService/StartContainer sandboxID=a22e1d52a296a877749d67bb8347e63d619ee2475777ed73628f6d33736e68d4 Jan 05 09:28:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:08.254363375Z" level=info msg="Started container" PID=3729 containerID=981d507021d63e67275789bb4c4c1625210c441f07a1e6819d475fb1eb066879 description=openshift-multus/multus-additional-cni-plugins-5fgxc/egress-router-binary-copy id=9ec47448-93f9-421a-a61e-072ddaec1ebc name=/runtime.v1.RuntimeService/StartContainer sandboxID=8ea5e7e9fb15db146766a6199273736a81c0e10f5714578c6922f99837137af4 Jan 05 09:28:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:08.276867256Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:08.277137867Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:08.277489511Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_e3ec878d-8bad-4254-879c-2fc80ccbc55f\"" Jan 05 09:28:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:08.355855318Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:08.355892173Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:08.355920102Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/multus\"" Jan 05 09:28:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:08.404176451Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:08.404472917Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:08.404593056Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_dc316480-6f54-41af-9926-89e3414d4400\"" Jan 05 09:28:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:08.404725312Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/egress-router\"" Jan 05 09:28:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:08.406469 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerStarted Data:981d507021d63e67275789bb4c4c1625210c441f07a1e6819d475fb1eb066879} Jan 05 09:28:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:08.410884 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-2r6nf" event=&{ID:aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f Type:ContainerStarted Data:bcf5ae9d6717d0c7922cf350cf9cecc4e5443de30e41d454fff49d818f809d14} Jan 05 09:28:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:08.413557 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-nxzr8" event=&{ID:8f88cbe5-1221-4e6f-a6c9-f22da152b43f Type:ContainerStarted Data:24638f270f12e2b431e69ccfc44265012351e11345c2e8b283d850d777824b59} Jan 05 09:28:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:08.444631 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-pbfzz" event=&{ID:f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7 Type:ContainerStarted Data:da19cbb49f674958c161a6b6f29875eb5d015a5a48e2c25147c41964dd2f2f0c} Jan 05 09:28:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:08.463480 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-xw6d2" event=&{ID:659b8bb4-81b5-49bf-8127-1d326f30ba2f Type:ContainerStarted Data:e40d040cc9cf721f176d342346bcd293f08fab396828b9dd6027d9f858fb489b} Jan 05 09:28:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:08.479946 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" event=&{ID:997b7ab4-939e-465c-9c7d-4a2ebe3a797a Type:ContainerStarted Data:0dde32b8bba42d827b7f9b8e7f3f00a32511565f8a3b5fbdb7328018feba4c3a} Jan 05 09:28:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:08.484136534Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:08.484454869Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:08.484478522Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_e3ec878d-8bad-4254-879c-2fc80ccbc55f\"" Jan 05 09:28:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:09.485657 2927 generic.go:296] "Generic (PLEG): container finished" podID=c0e47249-a5ff-460e-958d-53577387c154 containerID="981d507021d63e67275789bb4c4c1625210c441f07a1e6819d475fb1eb066879" exitCode=0 Jan 05 09:28:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:09.488263 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerDied Data:981d507021d63e67275789bb4c4c1625210c441f07a1e6819d475fb1eb066879} Jan 05 09:28:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:09.490740365Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d5fc161170fcf68900b2694039803a2e7f1b31eb913566139a41f91851f166d7" id=8ecb4c1c-2412-41bc-a353-a8c31ceb71b7 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:09.490973526Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:852cc803756e640141d6c3eb3e6607f2b0f7076ea77f9190b1f490288f40b3d8,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d5fc161170fcf68900b2694039803a2e7f1b31eb913566139a41f91851f166d7],Size_:574516705,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=8ecb4c1c-2412-41bc-a353-a8c31ceb71b7 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:09.494842399Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d5fc161170fcf68900b2694039803a2e7f1b31eb913566139a41f91851f166d7" id=3ba752b4-c196-4b25-b287-b08f13cd3c87 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:09.495214551Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:852cc803756e640141d6c3eb3e6607f2b0f7076ea77f9190b1f490288f40b3d8,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d5fc161170fcf68900b2694039803a2e7f1b31eb913566139a41f91851f166d7],Size_:574516705,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=3ba752b4-c196-4b25-b287-b08f13cd3c87 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:09.496030669Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/cni-plugins" id=2a48503e-1ffd-4b28-82f3-ea1dc29aa6c0 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:09.496232060Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:09.864510254Z" level=info msg="Created container 6de2d781e559a29d9b87b7b3efe271580ccd65b2604c19306d20462531bd445f: openshift-multus/multus-additional-cni-plugins-5fgxc/cni-plugins" id=2a48503e-1ffd-4b28-82f3-ea1dc29aa6c0 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:09.865154734Z" level=info msg="Starting container: 6de2d781e559a29d9b87b7b3efe271580ccd65b2604c19306d20462531bd445f" id=c0d6f911-156c-4306-bcb9-12c139ab9bed name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:28:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:09.879270509Z" level=info msg="Started container" PID=4084 containerID=6de2d781e559a29d9b87b7b3efe271580ccd65b2604c19306d20462531bd445f description=openshift-multus/multus-additional-cni-plugins-5fgxc/cni-plugins id=c0d6f911-156c-4306-bcb9-12c139ab9bed name=/runtime.v1.RuntimeService/StartContainer sandboxID=8ea5e7e9fb15db146766a6199273736a81c0e10f5714578c6922f99837137af4 Jan 05 09:28:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:09.891556804Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_c7b9c0ac-6d91-4e84-a261-0b5e83cb92bf\"" Jan 05 09:28:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:09.909192036Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:09.909340886Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.077725983Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/bandwidth\"" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.093081935Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.093122642Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.093139722Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/bridge\"" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.121772975Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.121939806Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.122209268Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/dhcp\"" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.136947469Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.137046461Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.137071819Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/firewall\"" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.148499081Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.148531333Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.148550494Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/host-device\"" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.157564578Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.157584766Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.157599539Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/host-local\"" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.166321210Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.166342591Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.166357725Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/ipvlan\"" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.176048110Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.176068778Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.176082962Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/loopback\"" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.183449752Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.183468497Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.183481664Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/macvlan\"" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.191220516Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.191239231Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.191253266Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/portmap\"" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.200290720Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.200309484Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.200323502Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/ptp\"" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.210126682Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.210145995Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.210159078Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/sbr\"" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.219437321Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.219455542Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.219469806Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/static\"" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.227465862Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.227484872Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.227497862Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/tuning\"" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.235555632Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.235573555Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.235586628Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/vlan\"" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.243215734Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.243233037Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.243245949Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/vrf\"" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.254797109Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.254816502Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.254831238Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_c7b9c0ac-6d91-4e84-a261-0b5e83cb92bf\"" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:10.489589 2927 generic.go:296] "Generic (PLEG): container finished" podID=c0e47249-a5ff-460e-958d-53577387c154 containerID="6de2d781e559a29d9b87b7b3efe271580ccd65b2604c19306d20462531bd445f" exitCode=0 Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:10.490249 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerDied Data:6de2d781e559a29d9b87b7b3efe271580ccd65b2604c19306d20462531bd445f} Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.490455483Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7ad3b9824ea195f38b6346262f0c609415c487ffaeb554725621a8327266da0f" id=c1496769-e7ff-49c9-92a4-1e7d138626f0 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.490717832Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:4c7cce1e79d06b3bc147fc59bd0a72991be97e767193f661db48974f3bb9ac30,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7ad3b9824ea195f38b6346262f0c609415c487ffaeb554725621a8327266da0f],Size_:336119748,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=c1496769-e7ff-49c9-92a4-1e7d138626f0 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.491337704Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7ad3b9824ea195f38b6346262f0c609415c487ffaeb554725621a8327266da0f" id=4e6e3f35-de3c-4f9d-bc17-f14c2d651005 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.491561219Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:4c7cce1e79d06b3bc147fc59bd0a72991be97e767193f661db48974f3bb9ac30,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7ad3b9824ea195f38b6346262f0c609415c487ffaeb554725621a8327266da0f],Size_:336119748,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=4e6e3f35-de3c-4f9d-bc17-f14c2d651005 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.492342507Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/bond-cni-plugin" id=76c095f4-363b-4d13-947d-c7a8c9e75f6b name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.492452741Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.847621637Z" level=info msg="Created container 1f49e803306e7666763e56e3f0c2ab5f2233bf09779217da27d7797559feafb5: openshift-multus/multus-additional-cni-plugins-5fgxc/bond-cni-plugin" id=76c095f4-363b-4d13-947d-c7a8c9e75f6b name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.848345153Z" level=info msg="Starting container: 1f49e803306e7666763e56e3f0c2ab5f2233bf09779217da27d7797559feafb5" id=ce78c6e8-0ae1-4471-b1bb-326804761c16 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.858388315Z" level=info msg="Started container" PID=4240 containerID=1f49e803306e7666763e56e3f0c2ab5f2233bf09779217da27d7797559feafb5 description=openshift-multus/multus-additional-cni-plugins-5fgxc/bond-cni-plugin id=ce78c6e8-0ae1-4471-b1bb-326804761c16 name=/runtime.v1.RuntimeService/StartContainer sandboxID=8ea5e7e9fb15db146766a6199273736a81c0e10f5714578c6922f99837137af4 Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.890949292Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_31f64da0-d970-4e36-a780-614012edb596\"" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.906041752Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.906065601Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.907440043Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/bond\"" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.945073732Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.945106599Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:10.945124054Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_31f64da0-d970-4e36-a780-614012edb596\"" Jan 05 09:28:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:11.493561 2927 generic.go:296] "Generic (PLEG): container finished" podID=c0e47249-a5ff-460e-958d-53577387c154 containerID="1f49e803306e7666763e56e3f0c2ab5f2233bf09779217da27d7797559feafb5" exitCode=0 Jan 05 09:28:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:11.493599 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerDied Data:1f49e803306e7666763e56e3f0c2ab5f2233bf09779217da27d7797559feafb5} Jan 05 09:28:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:11.494636889Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:a8bcc8cf428e222c7e6677fe5f7f05e79da2aed1ded851cad696d83c94860b4c" id=df44138f-f878-45be-9fb9-6c71013d114d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:11.495222882Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:34fa51fa29c1ebd2a3ad620255bb70edf89b0a96348f08efac3dc2ca3200602d,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:a8bcc8cf428e222c7e6677fe5f7f05e79da2aed1ded851cad696d83c94860b4c],Size_:275275974,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=df44138f-f878-45be-9fb9-6c71013d114d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:11.495878810Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:a8bcc8cf428e222c7e6677fe5f7f05e79da2aed1ded851cad696d83c94860b4c" id=c3db59f0-ec35-4170-a9f5-cdcd70299b34 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:11.496131455Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:34fa51fa29c1ebd2a3ad620255bb70edf89b0a96348f08efac3dc2ca3200602d,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:a8bcc8cf428e222c7e6677fe5f7f05e79da2aed1ded851cad696d83c94860b4c],Size_:275275974,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=c3db59f0-ec35-4170-a9f5-cdcd70299b34 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:11.496740354Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/routeoverride-cni" id=c901737f-0f29-41d3-a390-f0357515a4bd name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:11.496847820Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:11.760069326Z" level=info msg="Created container 746f03222421cec9e4e7e1a7353e0b9489c724a5443e2adcc147bb749e293191: openshift-multus/multus-additional-cni-plugins-5fgxc/routeoverride-cni" id=c901737f-0f29-41d3-a390-f0357515a4bd name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:11.760730319Z" level=info msg="Starting container: 746f03222421cec9e4e7e1a7353e0b9489c724a5443e2adcc147bb749e293191" id=601a4df9-7a33-49ca-988e-14e21bda8651 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:28:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:11.769630083Z" level=info msg="Started container" PID=4314 containerID=746f03222421cec9e4e7e1a7353e0b9489c724a5443e2adcc147bb749e293191 description=openshift-multus/multus-additional-cni-plugins-5fgxc/routeoverride-cni id=601a4df9-7a33-49ca-988e-14e21bda8651 name=/runtime.v1.RuntimeService/StartContainer sandboxID=8ea5e7e9fb15db146766a6199273736a81c0e10f5714578c6922f99837137af4 Jan 05 09:28:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:11.808703481Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_3c3faa50-e105-48e2-a9ad-0d9fdeba7f18\"" Jan 05 09:28:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:11.821285725Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:11.821435647Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:11.822125286Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/route-override\"" Jan 05 09:28:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:11.834444186Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:11.834471142Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:11.834488483Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_3c3faa50-e105-48e2-a9ad-0d9fdeba7f18\"" Jan 05 09:28:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:12.503898 2927 generic.go:296] "Generic (PLEG): container finished" podID=c0e47249-a5ff-460e-958d-53577387c154 containerID="746f03222421cec9e4e7e1a7353e0b9489c724a5443e2adcc147bb749e293191" exitCode=0 Jan 05 09:28:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:12.503958 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerDied Data:746f03222421cec9e4e7e1a7353e0b9489c724a5443e2adcc147bb749e293191} Jan 05 09:28:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:12.504969964Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451" id=59af836c-715e-4953-b897-7de308be5573 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:12.505197360Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:56f4df3c126a569448a12d5112b7c20e99e219b2b2fe6a04c4ba743396fd7751,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451],Size_:445620510,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=59af836c-715e-4953-b897-7de308be5573 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:12.506214938Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451" id=92517300-64e3-4eb3-b76d-3c50899d729c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:12.506505894Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:56f4df3c126a569448a12d5112b7c20e99e219b2b2fe6a04c4ba743396fd7751,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451],Size_:445620510,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=92517300-64e3-4eb3-b76d-3c50899d729c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:12.507135181Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/whereabouts-cni-bincopy" id=67a2c03a-e3fd-44df-9ee6-efea0c361161 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:12.507264395Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:12.715156917Z" level=info msg="Created container 3502daa90899b671d9c7a7bb8b57e7540cd3dddf9b2ddd79ab09724eb7e85a71: openshift-multus/multus-additional-cni-plugins-5fgxc/whereabouts-cni-bincopy" id=67a2c03a-e3fd-44df-9ee6-efea0c361161 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:12.715543813Z" level=info msg="Starting container: 3502daa90899b671d9c7a7bb8b57e7540cd3dddf9b2ddd79ab09724eb7e85a71" id=3325835c-71aa-4238-a4d0-f4a288077987 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:28:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:12.725044387Z" level=info msg="Started container" PID=4384 containerID=3502daa90899b671d9c7a7bb8b57e7540cd3dddf9b2ddd79ab09724eb7e85a71 description=openshift-multus/multus-additional-cni-plugins-5fgxc/whereabouts-cni-bincopy id=3325835c-71aa-4238-a4d0-f4a288077987 name=/runtime.v1.RuntimeService/StartContainer sandboxID=8ea5e7e9fb15db146766a6199273736a81c0e10f5714578c6922f99837137af4 Jan 05 09:28:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:12.728920176Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_d3fd2ae8-a996-4174-bf67-f9ac4bfa6627\"" Jan 05 09:28:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:12.747875318Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:12.748059961Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:12.827697172Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/whereabouts\"" Jan 05 09:28:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:12.843952953Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:12.844139204Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:12.844220173Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_d3fd2ae8-a996-4174-bf67-f9ac4bfa6627\"" Jan 05 09:28:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:13.507328 2927 generic.go:296] "Generic (PLEG): container finished" podID=c0e47249-a5ff-460e-958d-53577387c154 containerID="3502daa90899b671d9c7a7bb8b57e7540cd3dddf9b2ddd79ab09724eb7e85a71" exitCode=0 Jan 05 09:28:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:13.507835 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerDied Data:3502daa90899b671d9c7a7bb8b57e7540cd3dddf9b2ddd79ab09724eb7e85a71} Jan 05 09:28:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:13.508129449Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451" id=184d9830-a5c9-4395-bb90-2554b39b5960 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:13.508371434Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:56f4df3c126a569448a12d5112b7c20e99e219b2b2fe6a04c4ba743396fd7751,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451],Size_:445620510,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=184d9830-a5c9-4395-bb90-2554b39b5960 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:13.509211573Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451" id=8f650f22-e1cf-4cd0-bd5e-17c884db4953 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:13.509395269Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:56f4df3c126a569448a12d5112b7c20e99e219b2b2fe6a04c4ba743396fd7751,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451],Size_:445620510,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=8f650f22-e1cf-4cd0-bd5e-17c884db4953 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:13.510329807Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/whereabouts-cni" id=c520dcdc-f152-4481-85dc-7ccc52834a0e name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:13.510439404Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:13.693532865Z" level=info msg="Created container 8b883d7e785c3d77f13af2ea4a2ddce46b07766d3535eb50f487f6838eb87595: openshift-multus/multus-additional-cni-plugins-5fgxc/whereabouts-cni" id=c520dcdc-f152-4481-85dc-7ccc52834a0e name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:13.694048763Z" level=info msg="Starting container: 8b883d7e785c3d77f13af2ea4a2ddce46b07766d3535eb50f487f6838eb87595" id=ef6be2b0-b674-454f-92ff-ab1f33f9d2a5 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:28:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:13.701142048Z" level=info msg="Started container" PID=4457 containerID=8b883d7e785c3d77f13af2ea4a2ddce46b07766d3535eb50f487f6838eb87595 description=openshift-multus/multus-additional-cni-plugins-5fgxc/whereabouts-cni id=ef6be2b0-b674-454f-92ff-ab1f33f9d2a5 name=/runtime.v1.RuntimeService/StartContainer sandboxID=8ea5e7e9fb15db146766a6199273736a81c0e10f5714578c6922f99837137af4 Jan 05 09:28:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:14.512051 2927 generic.go:296] "Generic (PLEG): container finished" podID=c0e47249-a5ff-460e-958d-53577387c154 containerID="8b883d7e785c3d77f13af2ea4a2ddce46b07766d3535eb50f487f6838eb87595" exitCode=0 Jan 05 09:28:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:14.512089 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerDied Data:8b883d7e785c3d77f13af2ea4a2ddce46b07766d3535eb50f487f6838eb87595} Jan 05 09:28:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:14.512962786Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=b0217212-fd23-40a3-93c4-de6d1461f589 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:14.513201170Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b0217212-fd23-40a3-93c4-de6d1461f589 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:14.514141742Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=6cb49511-f311-42bb-a3ff-fc61ad8b26a8 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:14.514321723Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=6cb49511-f311-42bb-a3ff-fc61ad8b26a8 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:14.514933892Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/kube-multus-additional-cni-plugins" id=59b994d7-e2a3-40cb-ad88-8db81a6d3ed5 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:14.515055936Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:14.727179621Z" level=info msg="Created container 13b5ecec68198fbf9b177fedc00cf5f3cede3aa67da8574fca80b72643bb222f: openshift-multus/multus-additional-cni-plugins-5fgxc/kube-multus-additional-cni-plugins" id=59b994d7-e2a3-40cb-ad88-8db81a6d3ed5 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:14.727999180Z" level=info msg="Starting container: 13b5ecec68198fbf9b177fedc00cf5f3cede3aa67da8574fca80b72643bb222f" id=bcd65627-2ee0-4192-b7b2-0875659ac72f name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:28:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:14.738281418Z" level=info msg="Started container" PID=4522 containerID=13b5ecec68198fbf9b177fedc00cf5f3cede3aa67da8574fca80b72643bb222f description=openshift-multus/multus-additional-cni-plugins-5fgxc/kube-multus-additional-cni-plugins id=bcd65627-2ee0-4192-b7b2-0875659ac72f name=/runtime.v1.RuntimeService/StartContainer sandboxID=8ea5e7e9fb15db146766a6199273736a81c0e10f5714578c6922f99837137af4 Jan 05 09:28:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:15.191801 2927 patch_prober.go:29] interesting pod/alibaba-disk-csi-driver-node-5sqb4 container/csi-driver namespace/openshift-cluster-csi-drivers: Liveness probe status=failure output="Get \"http://10.0.99.113:10300/healthz\": dial tcp 10.0.99.113:10300: connect: connection refused" start-of-body= Jan 05 09:28:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:15.192031 2927 prober.go:114] "Probe failed" probeType="Liveness" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" podUID=1bb84a70-232c-4467-ac0b-647030155908 containerName="csi-driver" probeResult=failure output="Get \"http://10.0.99.113:10300/healthz\": dial tcp 10.0.99.113:10300: connect: connection refused" Jan 05 09:28:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:15.517190 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerStarted Data:13b5ecec68198fbf9b177fedc00cf5f3cede3aa67da8574fca80b72643bb222f} Jan 05 09:28:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:17.279048679Z" level=error msg="Failed to cleanup (probably retrying): failed to destroy network for pod sandbox k8s_thanos-querier-6f4f5fb7cf-rdm9l_openshift-monitoring_7c63c9be-d1ad-44b6-8b33-09af3ee314af_0(f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3): error removing pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/7c63c9be-d1ad-44b6-8b33-09af3ee314af]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" Jan 05 09:28:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:17.280015287Z" level=info msg="Got pod network &{Name:thanos-querier-6f4f5fb7cf-rdm9l Namespace:openshift-monitoring ID:60f4edc9298bb10efa1323e01ae8e14c10e33a612f4d624c4f4ce9aabba94896 UID:7c63c9be-d1ad-44b6-8b33-09af3ee314af NetNS:/var/run/netns/ea9064b2-7e7c-4016-be51-a63776aebfb5 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:28:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:17.280043467Z" level=info msg="Adding pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:28:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:19.810887512Z" level=info msg="cleanup sandbox network" Jan 05 09:28:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:25.192221 2927 patch_prober.go:29] interesting pod/alibaba-disk-csi-driver-node-5sqb4 container/csi-driver namespace/openshift-cluster-csi-drivers: Liveness probe status=failure output="Get \"http://10.0.99.113:10300/healthz\": dial tcp 10.0.99.113:10300: connect: connection refused" start-of-body= Jan 05 09:28:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:25.192275 2927 prober.go:114] "Probe failed" probeType="Liveness" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" podUID=1bb84a70-232c-4467-ac0b-647030155908 containerName="csi-driver" probeResult=failure output="Get \"http://10.0.99.113:10300/healthz\": dial tcp 10.0.99.113:10300: connect: connection refused" Jan 05 09:28:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:35.192082 2927 patch_prober.go:29] interesting pod/alibaba-disk-csi-driver-node-5sqb4 container/csi-driver namespace/openshift-cluster-csi-drivers: Liveness probe status=failure output="Get \"http://10.0.99.113:10300/healthz\": dial tcp 10.0.99.113:10300: connect: connection refused" start-of-body= Jan 05 09:28:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:35.192141 2927 prober.go:114] "Probe failed" probeType="Liveness" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" podUID=1bb84a70-232c-4467-ac0b-647030155908 containerName="csi-driver" probeResult=failure output="Get \"http://10.0.99.113:10300/healthz\": dial tcp 10.0.99.113:10300: connect: connection refused" Jan 05 09:28:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:42.027403 2927 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-multus/cni-sysctl-allowlist-ds-ccgld] Jan 05 09:28:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:42.027450 2927 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:28:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:42.153347 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/42d075ab-9d7c-4a5d-8c34-2d2b27434266-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-ccgld\" (UID: \"42d075ab-9d7c-4a5d-8c34-2d2b27434266\") " pod="openshift-multus/cni-sysctl-allowlist-ds-ccgld" Jan 05 09:28:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:42.153467 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zd8c\" (UniqueName: \"kubernetes.io/projected/42d075ab-9d7c-4a5d-8c34-2d2b27434266-kube-api-access-7zd8c\") pod \"cni-sysctl-allowlist-ds-ccgld\" (UID: \"42d075ab-9d7c-4a5d-8c34-2d2b27434266\") " pod="openshift-multus/cni-sysctl-allowlist-ds-ccgld" Jan 05 09:28:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:42.153615 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/42d075ab-9d7c-4a5d-8c34-2d2b27434266-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-ccgld\" (UID: \"42d075ab-9d7c-4a5d-8c34-2d2b27434266\") " pod="openshift-multus/cni-sysctl-allowlist-ds-ccgld" Jan 05 09:28:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:42.153676 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/42d075ab-9d7c-4a5d-8c34-2d2b27434266-ready\") pod \"cni-sysctl-allowlist-ds-ccgld\" (UID: \"42d075ab-9d7c-4a5d-8c34-2d2b27434266\") " pod="openshift-multus/cni-sysctl-allowlist-ds-ccgld" Jan 05 09:28:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:42.254485 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-7zd8c\" (UniqueName: \"kubernetes.io/projected/42d075ab-9d7c-4a5d-8c34-2d2b27434266-kube-api-access-7zd8c\") pod \"cni-sysctl-allowlist-ds-ccgld\" (UID: \"42d075ab-9d7c-4a5d-8c34-2d2b27434266\") " pod="openshift-multus/cni-sysctl-allowlist-ds-ccgld" Jan 05 09:28:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:42.254546 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/42d075ab-9d7c-4a5d-8c34-2d2b27434266-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-ccgld\" (UID: \"42d075ab-9d7c-4a5d-8c34-2d2b27434266\") " pod="openshift-multus/cni-sysctl-allowlist-ds-ccgld" Jan 05 09:28:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:42.254583 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/42d075ab-9d7c-4a5d-8c34-2d2b27434266-ready\") pod \"cni-sysctl-allowlist-ds-ccgld\" (UID: \"42d075ab-9d7c-4a5d-8c34-2d2b27434266\") " pod="openshift-multus/cni-sysctl-allowlist-ds-ccgld" Jan 05 09:28:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:42.254623 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/42d075ab-9d7c-4a5d-8c34-2d2b27434266-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-ccgld\" (UID: \"42d075ab-9d7c-4a5d-8c34-2d2b27434266\") " pod="openshift-multus/cni-sysctl-allowlist-ds-ccgld" Jan 05 09:28:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:42.254852 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/42d075ab-9d7c-4a5d-8c34-2d2b27434266-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-ccgld\" (UID: \"42d075ab-9d7c-4a5d-8c34-2d2b27434266\") " pod="openshift-multus/cni-sysctl-allowlist-ds-ccgld" Jan 05 09:28:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:42.255508 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/42d075ab-9d7c-4a5d-8c34-2d2b27434266-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-ccgld\" (UID: \"42d075ab-9d7c-4a5d-8c34-2d2b27434266\") " pod="openshift-multus/cni-sysctl-allowlist-ds-ccgld" Jan 05 09:28:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:42.255693 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/42d075ab-9d7c-4a5d-8c34-2d2b27434266-ready\") pod \"cni-sysctl-allowlist-ds-ccgld\" (UID: \"42d075ab-9d7c-4a5d-8c34-2d2b27434266\") " pod="openshift-multus/cni-sysctl-allowlist-ds-ccgld" Jan 05 09:28:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:42.283474 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zd8c\" (UniqueName: \"kubernetes.io/projected/42d075ab-9d7c-4a5d-8c34-2d2b27434266-kube-api-access-7zd8c\") pod \"cni-sysctl-allowlist-ds-ccgld\" (UID: \"42d075ab-9d7c-4a5d-8c34-2d2b27434266\") " pod="openshift-multus/cni-sysctl-allowlist-ds-ccgld" Jan 05 09:28:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:42.345946 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-ccgld" Jan 05 09:28:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:42.346446406Z" level=info msg="Running pod sandbox: openshift-multus/cni-sysctl-allowlist-ds-ccgld/POD" id=c34d6a47-595f-4cb5-a14e-b6448695ba61 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:42.346503472Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:42.363553146Z" level=info msg="Got pod network &{Name:cni-sysctl-allowlist-ds-ccgld Namespace:openshift-multus ID:6e0f82451fbe1ee8932e9ef9fb6a118b37c522bbdf8a12e24e0486a234053db8 UID:42d075ab-9d7c-4a5d-8c34-2d2b27434266 NetNS:/var/run/netns/bbe191ce-0210-43b0-a13f-ef8aa44d5a60 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:28:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:42.363715818Z" level=info msg="Adding pod openshift-multus_cni-sysctl-allowlist-ds-ccgld to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:28:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:45.191573 2927 patch_prober.go:29] interesting pod/alibaba-disk-csi-driver-node-5sqb4 container/csi-driver namespace/openshift-cluster-csi-drivers: Liveness probe status=failure output="Get \"http://10.0.99.113:10300/healthz\": dial tcp 10.0.99.113:10300: connect: connection refused" start-of-body= Jan 05 09:28:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:45.191632 2927 prober.go:114] "Probe failed" probeType="Liveness" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" podUID=1bb84a70-232c-4467-ac0b-647030155908 containerName="csi-driver" probeResult=failure output="Get \"http://10.0.99.113:10300/healthz\": dial tcp 10.0.99.113:10300: connect: connection refused" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.284021919Z" level=info msg="NetworkStart: stopping network for sandbox 74f272f4c50be5a640d85c29438b61a9a814b4b072afe86593507911550e6a03" id=84bf72ae-8e03-4616-a642-c3d68a94a275 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.284194362Z" level=info msg="Got pod network &{Name:prometheus-adapter-5894cc86c7-qhmbh Namespace:openshift-monitoring ID:74f272f4c50be5a640d85c29438b61a9a814b4b072afe86593507911550e6a03 UID:a40bf20d-67a4-45e7-b701-ca5a75376c1f NetNS:/var/run/netns/be37f359-70a2-4ca1-93a2-c054848967c1 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.284242385Z" level=error msg="error loading cached network config: network \"multus-cni-network\" not found in CNI cache" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.284256412Z" level=warning msg="falling back to loading from existing plugins on disk" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.284269167Z" level=info msg="Deleting pod openshift-monitoring_prometheus-adapter-5894cc86c7-qhmbh from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.301223274Z" level=info msg="NetworkStart: stopping network for sandbox ff598f9236d4f32150ffef17742e5f55d19d3b2adb4f7800e050a1ea04825617" id=b5f01488-c439-4bce-8eba-9abb8505890d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.301379771Z" level=info msg="Got pod network &{Name:prometheus-operator-admission-webhook-598468bb49-94dsn Namespace:openshift-monitoring ID:ff598f9236d4f32150ffef17742e5f55d19d3b2adb4f7800e050a1ea04825617 UID:88745ab7-efe7-42c0-a2e8-75591518333d NetNS:/var/run/netns/dcca133c-c11b-4381-a646-e2e95aca9314 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.301417005Z" level=error msg="error loading cached network config: network \"multus-cni-network\" not found in CNI cache" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.301431394Z" level=warning msg="falling back to loading from existing plugins on disk" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.301443169Z" level=info msg="Deleting pod openshift-monitoring_prometheus-operator-admission-webhook-598468bb49-94dsn from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.338826118Z" level=info msg="NetworkStart: stopping network for sandbox 1150fb360a7f91f41f9b7eb9840a298cfab29d821843d0d60688bc2927bdfece" id=3ef0c17d-bba5-488b-ad96-8e18ae386263 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.338993880Z" level=info msg="Got pod network &{Name:network-metrics-daemon-hq594 Namespace:openshift-multus ID:1150fb360a7f91f41f9b7eb9840a298cfab29d821843d0d60688bc2927bdfece UID:50152296-9489-4fa8-aa42-7858debe1c08 NetNS:/var/run/netns/e6314a5f-9fa7-4a60-ab76-6473053b4fa4 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.339030860Z" level=error msg="error loading cached network config: network \"multus-cni-network\" not found in CNI cache" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.339043408Z" level=warning msg="falling back to loading from existing plugins on disk" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.339053590Z" level=info msg="Deleting pod openshift-multus_network-metrics-daemon-hq594 from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.397851278Z" level=info msg="NetworkStart: stopping network for sandbox 8a255d5b5b10cf99b23616f05d7fb3c48ec24229e1f6cc5ce6484e928e0a86af" id=cea2f99d-429b-485e-9f18-1ae520031c8a name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.398019603Z" level=info msg="Got pod network &{Name:router-default-5cf6bbdff9-tnffs Namespace:openshift-ingress ID:8a255d5b5b10cf99b23616f05d7fb3c48ec24229e1f6cc5ce6484e928e0a86af UID:c3b10ff4-a243-4a4e-b27d-f613e258666d NetNS:/var/run/netns/e737e306-d06a-4291-bbea-c8e3872fce31 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.398058578Z" level=error msg="error loading cached network config: network \"multus-cni-network\" not found in CNI cache" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.398070975Z" level=warning msg="falling back to loading from existing plugins on disk" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.398082898Z" level=info msg="Deleting pod openshift-ingress_router-default-5cf6bbdff9-tnffs from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.416056643Z" level=info msg="NetworkStart: stopping network for sandbox 52b90175ef0a6f25e9a947f029d3b67d80d2c31734082f3c71e38fa2ba264d89" id=0e4a725d-2cd9-4f62-8e17-541b1dbb4f78 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.416209419Z" level=info msg="Got pod network &{Name:prometheus-k8s-0 Namespace:openshift-monitoring ID:52b90175ef0a6f25e9a947f029d3b67d80d2c31734082f3c71e38fa2ba264d89 UID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 NetNS:/var/run/netns/08082c34-5052-4e18-a998-520794be8216 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.416243222Z" level=error msg="error loading cached network config: network \"multus-cni-network\" not found in CNI cache" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.416256073Z" level=warning msg="falling back to loading from existing plugins on disk" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.416267318Z" level=info msg="Deleting pod openshift-monitoring_prometheus-k8s-0 from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.432707929Z" level=info msg="NetworkStart: stopping network for sandbox ed0ee704407b6914733819122eb9fc60ca3996f0c1b935170d5903732b869347" id=bc650832-6cc0-4148-bdb8-05b5ae4deaf1 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.432899191Z" level=info msg="Got pod network &{Name:alertmanager-main-0 Namespace:openshift-monitoring ID:ed0ee704407b6914733819122eb9fc60ca3996f0c1b935170d5903732b869347 UID:05c9a6ea-140a-49a2-bde5-f220b05aa252 NetNS:/var/run/netns/e5957241-af0b-4350-a062-91e5162d6a40 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.432948630Z" level=error msg="error loading cached network config: network \"multus-cni-network\" not found in CNI cache" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.432961345Z" level=warning msg="falling back to loading from existing plugins on disk" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.432971966Z" level=info msg="Deleting pod openshift-monitoring_alertmanager-main-0 from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.524005581Z" level=info msg="NetworkStart: stopping network for sandbox 2c2a604331152b239cd95403d2bb266e2ca816b6a0414973001dc448b2d2b416" id=606b9627-db1b-4a3f-874b-b462a6d88dd8 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.524166284Z" level=info msg="Got pod network &{Name:ingress-canary-8xrbh Namespace:openshift-ingress-canary ID:2c2a604331152b239cd95403d2bb266e2ca816b6a0414973001dc448b2d2b416 UID:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748 NetNS:/var/run/netns/28d99893-e7fa-4b86-9cf0-5e450f5457d2 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.524211882Z" level=error msg="error loading cached network config: network \"multus-cni-network\" not found in CNI cache" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.524226361Z" level=warning msg="falling back to loading from existing plugins on disk" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.524237647Z" level=info msg="Deleting pod openshift-ingress-canary_ingress-canary-8xrbh from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.595945722Z" level=info msg="NetworkStart: stopping network for sandbox a16c8f8a13d2354862769ea877c06e032be034fdb676b065e5595fbceae680cf" id=64b71f4f-3284-4db9-a02a-4cd2ff5dd755 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.596350758Z" level=info msg="Got pod network &{Name:telemeter-client-5598c645c4-mftdv Namespace:openshift-monitoring ID:a16c8f8a13d2354862769ea877c06e032be034fdb676b065e5595fbceae680cf UID:fc45e847-8f55-4c48-8c48-4bfee2419dca NetNS:/var/run/netns/96790582-8aca-42b0-a796-fa0aa2193bfa Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.596394940Z" level=error msg="error loading cached network config: network \"multus-cni-network\" not found in CNI cache" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.596416109Z" level=warning msg="falling back to loading from existing plugins on disk" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.596427577Z" level=info msg="Deleting pod openshift-monitoring_telemeter-client-5598c645c4-mftdv from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.607719236Z" level=info msg="NetworkStart: stopping network for sandbox dcf02c2c8b254c0ba075cbb4447c97111fedbd18176c9b10dcf0b8e0fa4c2b1d" id=e1a3ca46-101b-43fa-9721-c16b2c482369 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.607878257Z" level=info msg="Got pod network &{Name:dns-default-2pfzf Namespace:openshift-dns ID:dcf02c2c8b254c0ba075cbb4447c97111fedbd18176c9b10dcf0b8e0fa4c2b1d UID:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 NetNS:/var/run/netns/f79bd825-de4b-4a51-9fe9-dc217b8ab2c4 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.607928085Z" level=error msg="error loading cached network config: network \"multus-cni-network\" not found in CNI cache" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.607943206Z" level=warning msg="falling back to loading from existing plugins on disk" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.607955198Z" level=info msg="Deleting pod openshift-dns_dns-default-2pfzf from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.693029603Z" level=info msg="NetworkStart: stopping network for sandbox a937bd9597c90954ad3041a6e3670bd2dfc453dbbd7b4d1b6d48e35dae7b83c0" id=0755a791-7501-4417-934f-53470d89bb87 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.693189133Z" level=info msg="Got pod network &{Name:kube-state-metrics-6798b548b9-xl88q Namespace:openshift-monitoring ID:a937bd9597c90954ad3041a6e3670bd2dfc453dbbd7b4d1b6d48e35dae7b83c0 UID:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e NetNS:/var/run/netns/7dcc1407-51e3-46a5-8bb4-1903ac68eef7 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.693230326Z" level=error msg="error loading cached network config: network \"multus-cni-network\" not found in CNI cache" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.693241644Z" level=warning msg="falling back to loading from existing plugins on disk" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.693253274Z" level=info msg="Deleting pod openshift-monitoring_kube-state-metrics-6798b548b9-xl88q from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.749959703Z" level=info msg="NetworkStart: stopping network for sandbox 68c40ad6338d492b4aa5a2126d3afd81d13c48ead9d59f5b59bd6ce18b97e366" id=96524ba2-8daa-4a74-ab1c-222f3b9d9054 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.750119712Z" level=info msg="Got pod network &{Name:network-check-target-xmq2g Namespace:openshift-network-diagnostics ID:68c40ad6338d492b4aa5a2126d3afd81d13c48ead9d59f5b59bd6ce18b97e366 UID:f364a949-09d7-4cfb-83ff-e532b822a557 NetNS:/var/run/netns/eee4c6a8-243a-4fce-8d4b-4475692c8e09 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.750154685Z" level=error msg="error loading cached network config: network \"multus-cni-network\" not found in CNI cache" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.750167003Z" level=warning msg="falling back to loading from existing plugins on disk" Jan 05 09:28:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:50.750177231Z" level=info msg="Deleting pod openshift-network-diagnostics_network-check-target-xmq2g from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:28:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:51.357955178Z" level=info msg="NetworkStart: stopping network for sandbox afb580e701142b75f062c265b00883e2d5567b37000d823ade7ce8eb8a057b02" id=ff04d155-666b-4a8f-ba4b-1d1da011fa66 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:28:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:51.358715296Z" level=info msg="Got pod network &{Name:openshift-state-metrics-6f5dcdf79b-wq256 Namespace:openshift-monitoring ID:afb580e701142b75f062c265b00883e2d5567b37000d823ade7ce8eb8a057b02 UID:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 NetNS:/var/run/netns/66f76849-1ec6-4745-9cd9-8c97630f69f8 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:28:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:51.358885980Z" level=error msg="error loading cached network config: network \"multus-cni-network\" not found in CNI cache" Jan 05 09:28:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:51.358992153Z" level=warning msg="falling back to loading from existing plugins on disk" Jan 05 09:28:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:51.359055727Z" level=info msg="Deleting pod openshift-monitoring_openshift-state-metrics-6f5dcdf79b-wq256 from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:28:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:55.192131 2927 patch_prober.go:29] interesting pod/alibaba-disk-csi-driver-node-5sqb4 container/csi-driver namespace/openshift-cluster-csi-drivers: Liveness probe status=failure output="Get \"http://10.0.99.113:10300/healthz\": dial tcp 10.0.99.113:10300: connect: connection refused" start-of-body= Jan 05 09:28:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:55.192196 2927 prober.go:114] "Probe failed" probeType="Liveness" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" podUID=1bb84a70-232c-4467-ac0b-647030155908 containerName="csi-driver" probeResult=failure output="Get \"http://10.0.99.113:10300/healthz\": dial tcp 10.0.99.113:10300: connect: connection refused" Jan 05 09:28:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:55.192231 2927 kubelet.go:2229] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:28:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:55.193474 2927 kuberuntime_manager.go:649] "Message for Container of pod" containerName="csi-driver" containerStatusID={Type:cri-o ID:eb2903e092d65e9ea578f6d3048aa2bcf5b3ed144f88c5b61a8a443631ae7f6b} pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" containerMessage="Container csi-driver failed liveness probe, will be restarted" Jan 05 09:28:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:55.193855 2927 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" podUID=1bb84a70-232c-4467-ac0b-647030155908 containerName="csi-driver" containerID="cri-o://eb2903e092d65e9ea578f6d3048aa2bcf5b3ed144f88c5b61a8a443631ae7f6b" gracePeriod=30 Jan 05 09:28:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:55.194331888Z" level=info msg="Stopping container: eb2903e092d65e9ea578f6d3048aa2bcf5b3ed144f88c5b61a8a443631ae7f6b (timeout: 30s)" id=970310da-371e-410b-958c-2780a5f3db3c name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:28:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:55.589203 2927 generic.go:296] "Generic (PLEG): container finished" podID=1bb84a70-232c-4467-ac0b-647030155908 containerID="eb2903e092d65e9ea578f6d3048aa2bcf5b3ed144f88c5b61a8a443631ae7f6b" exitCode=143 Jan 05 09:28:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:55.589247 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" event=&{ID:1bb84a70-232c-4467-ac0b-647030155908 Type:ContainerDied Data:eb2903e092d65e9ea578f6d3048aa2bcf5b3ed144f88c5b61a8a443631ae7f6b} Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.368088388Z" level=info msg="Stopped container eb2903e092d65e9ea578f6d3048aa2bcf5b3ed144f88c5b61a8a443631ae7f6b: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-driver" id=970310da-371e-410b-958c-2780a5f3db3c name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.369604336Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7" id=5dc55f17-5885-420e-832e-55f9fc3858bc name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.373511606Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:6e78a44a973f2298da6f51c1418094403ee8c6c2b57bde9d3285b9e2ac2ba129,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7],Size_:397889589,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=5dc55f17-5885-420e-832e-55f9fc3858bc name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.374244788Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7" id=26baccdd-6baa-4aee-970a-1171770ac558 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.378015113Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:6e78a44a973f2298da6f51c1418094403ee8c6c2b57bde9d3285b9e2ac2ba129,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7],Size_:397889589,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=26baccdd-6baa-4aee-970a-1171770ac558 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.378886364Z" level=info msg="Creating container: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-driver" id=8a4bc5bd-54d8-4d9b-b965-33192d395168 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.379100355Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.415355316Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:96010fa3364bd5030d4445bbc80df5c8ec18d2b5579f23f71235bcda47b38ca6" id=c719d34b-89a5-4d84-adde-2dcf5972639b name=/runtime.v1.ImageService/PullImage Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.416626984Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:96010fa3364bd5030d4445bbc80df5c8ec18d2b5579f23f71235bcda47b38ca6" id=a30de7d8-dab8-42b4-8d00-fe025b60b849 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.424286074Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:e79b08638906e1c69cc20601e5eaf395f9830e1112e2dea73cbf10739b831dcc,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:96010fa3364bd5030d4445bbc80df5c8ec18d2b5579f23f71235bcda47b38ca6],Size_:1128011356,Uid:nil,Username:root,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=a30de7d8-dab8-42b4-8d00-fe025b60b849 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.425698317Z" level=info msg="Creating container: openshift-ovn-kubernetes/ovnkube-node-rp9bd/ovn-controller" id=6e885467-f0ae-4040-b4c3-9c579bc5c63c name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.426087896Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.527626327Z" level=info msg="Created container f254ffcbf63c78d47c512efd7a77565e19e122ecfc052bf42a2fa62c80045a60: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-driver" id=8a4bc5bd-54d8-4d9b-b965-33192d395168 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.528285305Z" level=info msg="Starting container: f254ffcbf63c78d47c512efd7a77565e19e122ecfc052bf42a2fa62c80045a60" id=8c73e3af-7478-4563-b5e8-39f554ce71ba name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.545066609Z" level=info msg="Started container" PID=4971 containerID=f254ffcbf63c78d47c512efd7a77565e19e122ecfc052bf42a2fa62c80045a60 description=openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-driver id=8c73e3af-7478-4563-b5e8-39f554ce71ba name=/runtime.v1.RuntimeService/StartContainer sandboxID=23c255a61eb7c200b97ccd2916bb330a59475bd2d3e145171f1e9f94a9c27e2e Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.555524453Z" level=info msg="Created container 2a002f48611de0f066f753f27aa97988bb851c0533c009a991efe9a325cfa713: openshift-ovn-kubernetes/ovnkube-node-rp9bd/ovn-controller" id=6e885467-f0ae-4040-b4c3-9c579bc5c63c name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.556274679Z" level=info msg="Starting container: 2a002f48611de0f066f753f27aa97988bb851c0533c009a991efe9a325cfa713" id=5017ea8f-9730-4093-8292-b0f2e9c7f0c1 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.577518097Z" level=info msg="Started container" PID=4981 containerID=2a002f48611de0f066f753f27aa97988bb851c0533c009a991efe9a325cfa713 description=openshift-ovn-kubernetes/ovnkube-node-rp9bd/ovn-controller id=5017ea8f-9730-4093-8292-b0f2e9c7f0c1 name=/runtime.v1.RuntimeService/StartContainer sandboxID=fd9deb131d1f752a07a6704193bcd1cf1dab19ff2e6439be2d911f94c338da51 Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:56.592725 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" event=&{ID:aac1c4d4-278d-4a53-b2f7-a75ba43c2d18 Type:ContainerStarted Data:2a002f48611de0f066f753f27aa97988bb851c0533c009a991efe9a325cfa713} Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.593865255Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:96010fa3364bd5030d4445bbc80df5c8ec18d2b5579f23f71235bcda47b38ca6" id=08c9f62d-ba31-447a-ab32-d7eeb37622ac name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.598085642Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:e79b08638906e1c69cc20601e5eaf395f9830e1112e2dea73cbf10739b831dcc,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:96010fa3364bd5030d4445bbc80df5c8ec18d2b5579f23f71235bcda47b38ca6],Size_:1128011356,Uid:nil,Username:root,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=08c9f62d-ba31-447a-ab32-d7eeb37622ac name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:56.599202 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" event=&{ID:1bb84a70-232c-4467-ac0b-647030155908 Type:ContainerStarted Data:f254ffcbf63c78d47c512efd7a77565e19e122ecfc052bf42a2fa62c80045a60} Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.600883531Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:96010fa3364bd5030d4445bbc80df5c8ec18d2b5579f23f71235bcda47b38ca6" id=f5c5df71-0bf2-445f-8a4c-e01183e73520 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.603230382Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:e79b08638906e1c69cc20601e5eaf395f9830e1112e2dea73cbf10739b831dcc,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:96010fa3364bd5030d4445bbc80df5c8ec18d2b5579f23f71235bcda47b38ca6],Size_:1128011356,Uid:nil,Username:root,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=f5c5df71-0bf2-445f-8a4c-e01183e73520 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.605438472Z" level=info msg="Creating container: openshift-ovn-kubernetes/ovnkube-node-rp9bd/ovn-acl-logging" id=6dda1c51-4a1a-41ab-af62-b4509f83c443 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.605552783Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.728710696Z" level=info msg="Created container a810bb6ca3335cd24a04949ad82533875c7fd3127e211ae9ebec6cea2c3dc045: openshift-ovn-kubernetes/ovnkube-node-rp9bd/ovn-acl-logging" id=6dda1c51-4a1a-41ab-af62-b4509f83c443 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.729532353Z" level=info msg="Starting container: a810bb6ca3335cd24a04949ad82533875c7fd3127e211ae9ebec6cea2c3dc045" id=f75d1cee-d563-4a53-af2c-ddaa02aa8578 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.736729763Z" level=info msg="Started container" PID=5046 containerID=a810bb6ca3335cd24a04949ad82533875c7fd3127e211ae9ebec6cea2c3dc045 description=openshift-ovn-kubernetes/ovnkube-node-rp9bd/ovn-acl-logging id=f75d1cee-d563-4a53-af2c-ddaa02aa8578 name=/runtime.v1.RuntimeService/StartContainer sandboxID=fd9deb131d1f752a07a6704193bcd1cf1dab19ff2e6439be2d911f94c338da51 Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.747890965Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=342200fd-2e7c-45d5-bdf6-28af89a0b2dc name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.748177860Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=342200fd-2e7c-45d5-bdf6-28af89a0b2dc name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.749062717Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=0bd301cc-fc48-45ca-8bd0-120595d44422 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.749471780Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=0bd301cc-fc48-45ca-8bd0-120595d44422 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.750691214Z" level=info msg="Creating container: openshift-ovn-kubernetes/ovnkube-node-rp9bd/kube-rbac-proxy" id=eb20b0f9-4c27-4a75-9616-a530d8edae83 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.750954629Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.852182938Z" level=info msg="Created container a942da66e3e5e6ad4915085345ab0c7d57f83d8c68551251fe4b168376e26596: openshift-ovn-kubernetes/ovnkube-node-rp9bd/kube-rbac-proxy" id=eb20b0f9-4c27-4a75-9616-a530d8edae83 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.852770327Z" level=info msg="Starting container: a942da66e3e5e6ad4915085345ab0c7d57f83d8c68551251fe4b168376e26596" id=2b1c41a1-c013-4447-a056-cb2dfed0a56a name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.861928538Z" level=info msg="Started container" PID=5087 containerID=a942da66e3e5e6ad4915085345ab0c7d57f83d8c68551251fe4b168376e26596 description=openshift-ovn-kubernetes/ovnkube-node-rp9bd/kube-rbac-proxy id=2b1c41a1-c013-4447-a056-cb2dfed0a56a name=/runtime.v1.RuntimeService/StartContainer sandboxID=fd9deb131d1f752a07a6704193bcd1cf1dab19ff2e6439be2d911f94c338da51 Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.869828685Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=3873673c-cf5e-4fa2-9048-871ceae5b33a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.870069941Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=3873673c-cf5e-4fa2-9048-871ceae5b33a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.870756231Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=b9dbbd91-2e83-4743-b80a-5b1008ad4b4e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.870994114Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b9dbbd91-2e83-4743-b80a-5b1008ad4b4e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.871745116Z" level=info msg="Creating container: openshift-ovn-kubernetes/ovnkube-node-rp9bd/kube-rbac-proxy-ovn-metrics" id=6a2cf9e1-862a-4c30-97cd-bd5b175196ea name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.871876200Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.994093423Z" level=info msg="Created container d0db0990bab100d4dde6c13a68393ef82c94744cd6a2a6d42cf1f7fc68e5a389: openshift-ovn-kubernetes/ovnkube-node-rp9bd/kube-rbac-proxy-ovn-metrics" id=6a2cf9e1-862a-4c30-97cd-bd5b175196ea name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:56.994481558Z" level=info msg="Starting container: d0db0990bab100d4dde6c13a68393ef82c94744cd6a2a6d42cf1f7fc68e5a389" id=a16f56c0-637f-49a3-ae96-8c42232d4394 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:57.004670567Z" level=info msg="Started container" PID=5132 containerID=d0db0990bab100d4dde6c13a68393ef82c94744cd6a2a6d42cf1f7fc68e5a389 description=openshift-ovn-kubernetes/ovnkube-node-rp9bd/kube-rbac-proxy-ovn-metrics id=a16f56c0-637f-49a3-ae96-8c42232d4394 name=/runtime.v1.RuntimeService/StartContainer sandboxID=fd9deb131d1f752a07a6704193bcd1cf1dab19ff2e6439be2d911f94c338da51 Jan 05 09:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:57.014425556Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:96010fa3364bd5030d4445bbc80df5c8ec18d2b5579f23f71235bcda47b38ca6" id=24898761-bc1a-4f88-8f9e-20600acec616 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:57.017113403Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:e79b08638906e1c69cc20601e5eaf395f9830e1112e2dea73cbf10739b831dcc,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:96010fa3364bd5030d4445bbc80df5c8ec18d2b5579f23f71235bcda47b38ca6],Size_:1128011356,Uid:nil,Username:root,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=24898761-bc1a-4f88-8f9e-20600acec616 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:57.018055363Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:96010fa3364bd5030d4445bbc80df5c8ec18d2b5579f23f71235bcda47b38ca6" id=a86efb8f-937b-4ec7-bed7-f8b59dddc914 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:57.020459377Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:e79b08638906e1c69cc20601e5eaf395f9830e1112e2dea73cbf10739b831dcc,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:96010fa3364bd5030d4445bbc80df5c8ec18d2b5579f23f71235bcda47b38ca6],Size_:1128011356,Uid:nil,Username:root,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=a86efb8f-937b-4ec7-bed7-f8b59dddc914 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:57.022204202Z" level=info msg="Creating container: openshift-ovn-kubernetes/ovnkube-node-rp9bd/ovnkube-node" id=7944d7cd-a86f-4797-a3f5-47653b53240b name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:57.022311047Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:57.106929811Z" level=info msg="Created container 23a50821ca44e2b87851e25e4b341565d308519d780b1b4bb88614b86cab434b: openshift-ovn-kubernetes/ovnkube-node-rp9bd/ovnkube-node" id=7944d7cd-a86f-4797-a3f5-47653b53240b name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:57.107317186Z" level=info msg="Starting container: 23a50821ca44e2b87851e25e4b341565d308519d780b1b4bb88614b86cab434b" id=6fde541b-bb81-444c-b040-d394d3f0a7fe name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:57.114980387Z" level=info msg="Started container" PID=5179 containerID=23a50821ca44e2b87851e25e4b341565d308519d780b1b4bb88614b86cab434b description=openshift-ovn-kubernetes/ovnkube-node-rp9bd/ovnkube-node id=6fde541b-bb81-444c-b040-d394d3f0a7fe name=/runtime.v1.RuntimeService/StartContainer sandboxID=fd9deb131d1f752a07a6704193bcd1cf1dab19ff2e6439be2d911f94c338da51 Jan 05 09:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:57.117508582Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/ovn-k8s-cni-overlay\"" Jan 05 09:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:57.132916632Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:57.133062915Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:57.133152913Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/ovn-k8s-cni-overlay\"" Jan 05 09:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:57.147467879Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:57.147634939Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:57.147756045Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/ovn-k8s-cni-overlay\"" Jan 05 09:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:57.165286245Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:57.165472831Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:57.165575359Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/ovn-k8s-cni-overlay\"" Jan 05 09:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:57.183058555Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:57.183085558Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:57.183104562Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/ovn-k8s-cni-overlay\"" Jan 05 09:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:57.199968627Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:28:57.199995439Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:57.602570 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" event=&{ID:aac1c4d4-278d-4a53-b2f7-a75ba43c2d18 Type:ContainerStarted Data:23a50821ca44e2b87851e25e4b341565d308519d780b1b4bb88614b86cab434b} Jan 05 09:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:57.602874 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" event=&{ID:aac1c4d4-278d-4a53-b2f7-a75ba43c2d18 Type:ContainerStarted Data:d0db0990bab100d4dde6c13a68393ef82c94744cd6a2a6d42cf1f7fc68e5a389} Jan 05 09:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:57.602891 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" event=&{ID:aac1c4d4-278d-4a53-b2f7-a75ba43c2d18 Type:ContainerStarted Data:a942da66e3e5e6ad4915085345ab0c7d57f83d8c68551251fe4b168376e26596} Jan 05 09:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:57.602920 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" event=&{ID:aac1c4d4-278d-4a53-b2f7-a75ba43c2d18 Type:ContainerStarted Data:a810bb6ca3335cd24a04949ad82533875c7fd3127e211ae9ebec6cea2c3dc045} Jan 05 09:28:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:28:57.604033 2927 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:29:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:02.295322855Z" level=info msg="NetworkStart: stopping network for sandbox 60f4edc9298bb10efa1323e01ae8e14c10e33a612f4d624c4f4ce9aabba94896" id=510e0b76-941b-4df9-96cf-95b4608a89ed name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:02.295494297Z" level=info msg="Got pod network &{Name:thanos-querier-6f4f5fb7cf-rdm9l Namespace:openshift-monitoring ID:60f4edc9298bb10efa1323e01ae8e14c10e33a612f4d624c4f4ce9aabba94896 UID:7c63c9be-d1ad-44b6-8b33-09af3ee314af NetNS:/var/run/netns/ea9064b2-7e7c-4016-be51-a63776aebfb5 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:02.295531561Z" level=error msg="error loading cached network config: network \"multus-cni-network\" not found in CNI cache" Jan 05 09:29:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:02.295542159Z" level=warning msg="falling back to loading from existing plugins on disk" Jan 05 09:29:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:02.295552423Z" level=info msg="Deleting pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:29:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:05.626271 2927 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" podUID=aac1c4d4-278d-4a53-b2f7-a75ba43c2d18 containerName="ovnkube-node" probeResult=failure output="" Jan 05 09:29:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:08.651396 2927 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh] Jan 05 09:29:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:08.676043 2927 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-dns/dns-default-2pfzf] Jan 05 09:29:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:08.703923 2927 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256] Jan 05 09:29:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:08.816460 2927 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/prometheus-k8s-0] Jan 05 09:29:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:09.017598 2927 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-multus/cni-sysctl-allowlist-ds-ccgld] Jan 05 09:29:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:09.132871 2927 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-network-diagnostics/network-check-target-xmq2g] Jan 05 09:29:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:10.057097 2927 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/alertmanager-main-0] Jan 05 09:29:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:10.074209 2927 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-ingress/router-default-5cf6bbdff9-tnffs] Jan 05 09:29:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:10.150732 2927 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/telemeter-client-5598c645c4-mftdv] Jan 05 09:29:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:10.417304 2927 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-multus/network-metrics-daemon-hq594] Jan 05 09:29:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:10.443315 2927 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/kube-state-metrics-6798b548b9-xl88q] Jan 05 09:29:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:10.522040 2927 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-ingress-canary/ingress-canary-8xrbh] Jan 05 09:29:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:10.626080 2927 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" podUID=aac1c4d4-278d-4a53-b2f7-a75ba43c2d18 containerName="ovnkube-node" probeResult=failure output="" Jan 05 09:29:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:10.804964 2927 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn] Jan 05 09:29:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:10.932173 2927 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l] Jan 05 09:29:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:15.191735 2927 patch_prober.go:29] interesting pod/alibaba-disk-csi-driver-node-5sqb4 container/csi-driver namespace/openshift-cluster-csi-drivers: Liveness probe status=failure output="Get \"http://10.0.99.113:10300/healthz\": dial tcp 10.0.99.113:10300: connect: connection refused" start-of-body= Jan 05 09:29:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:15.191789 2927 prober.go:114] "Probe failed" probeType="Liveness" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" podUID=1bb84a70-232c-4467-ac0b-647030155908 containerName="csi-driver" probeResult=failure output="Get \"http://10.0.99.113:10300/healthz\": dial tcp 10.0.99.113:10300: connect: connection refused" Jan 05 09:29:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:15.647262 2927 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" podUID=aac1c4d4-278d-4a53-b2f7-a75ba43c2d18 containerName="ovnkube-node" probeResult=failure output="" Jan 05 09:29:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:17.175382322Z" level=info msg="CNI monitoring event REMOVE \"/etc/kubernetes/cni/net.d/00-multus.conf\"" Jan 05 09:29:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:17.192888330Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:29:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:17.192967159Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:29:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:17.192989315Z" level=info msg="CNI monitoring event CREATE \"/etc/kubernetes/cni/net.d/00-multus.conf\"" Jan 05 09:29:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:17.201970049Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:29:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:17.201993898Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:29:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:17.202012308Z" level=info msg="CNI monitoring event WRITE \"/etc/kubernetes/cni/net.d/00-multus.conf\"" Jan 05 09:29:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:17.209327673Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:29:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:17.209347522Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:29:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:17.209361007Z" level=info msg="CNI monitoring event CHMOD \"/etc/kubernetes/cni/net.d/00-multus.conf\"" Jan 05 09:29:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:20.625490 2927 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" Jan 05 09:29:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:25.191813 2927 patch_prober.go:29] interesting pod/alibaba-disk-csi-driver-node-5sqb4 container/csi-driver namespace/openshift-cluster-csi-drivers: Liveness probe status=failure output="Get \"http://10.0.99.113:10300/healthz\": dial tcp 10.0.99.113:10300: connect: connection refused" start-of-body= Jan 05 09:29:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:25.191910 2927 prober.go:114] "Probe failed" probeType="Liveness" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" podUID=1bb84a70-232c-4467-ac0b-647030155908 containerName="csi-driver" probeResult=failure output="Get \"http://10.0.99.113:10300/healthz\": dial tcp 10.0.99.113:10300: connect: connection refused" Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:27.386126459Z" level=info msg="NetworkStart: stopping network for sandbox 6e0f82451fbe1ee8932e9ef9fb6a118b37c522bbdf8a12e24e0486a234053db8" id=c34d6a47-595f-4cb5-a14e-b6448695ba61 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:27.386345934Z" level=info msg="Got pod network &{Name:cni-sysctl-allowlist-ds-ccgld Namespace:openshift-multus ID:6e0f82451fbe1ee8932e9ef9fb6a118b37c522bbdf8a12e24e0486a234053db8 UID:42d075ab-9d7c-4a5d-8c34-2d2b27434266 NetNS:/var/run/netns/bbe191ce-0210-43b0-a13f-ef8aa44d5a60 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:27.386383951Z" level=error msg="error loading cached network config: network \"multus-cni-network\" not found in CNI cache" Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:27.386397337Z" level=warning msg="falling back to loading from existing plugins on disk" Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:27.386408496Z" level=info msg="Deleting pod openshift-multus_cni-sysctl-allowlist-ds-ccgld from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: 2023-01-05T09:29:27Z [verbose] Del: openshift-multus:cni-sysctl-allowlist-ds-ccgld:42d075ab-9d7c-4a5d-8c34-2d2b27434266:ovn-kubernetes:eth0 {"cniVersion":"0.4.0","dns":{},"ipam":{},"logFile":"/var/log/ovn-kubernetes/ovn-k8s-cni-overlay.log","logLevel":"4","logfile-maxage":5,"logfile-maxbackups":5,"logfile-maxsize":100,"name":"ovn-kubernetes","type":"ovn-k8s-cni-overlay"} Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: I0105 09:29:27.538990 6030 ovs.go:90] Maximum command line arguments set to: 191102 Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:27.572664443Z" level=info msg="runSandbox: cleaning up namespaces after failing to run sandbox 6e0f82451fbe1ee8932e9ef9fb6a118b37c522bbdf8a12e24e0486a234053db8" id=c34d6a47-595f-4cb5-a14e-b6448695ba61 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:27.600051095Z" level=info msg="runSandbox: deleting pod ID 6e0f82451fbe1ee8932e9ef9fb6a118b37c522bbdf8a12e24e0486a234053db8 from idIndex" id=c34d6a47-595f-4cb5-a14e-b6448695ba61 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:27.600097167Z" level=info msg="runSandbox: removing pod sandbox 6e0f82451fbe1ee8932e9ef9fb6a118b37c522bbdf8a12e24e0486a234053db8" id=c34d6a47-595f-4cb5-a14e-b6448695ba61 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:27.600121850Z" level=info msg="runSandbox: deleting container ID from idIndex for sandbox 6e0f82451fbe1ee8932e9ef9fb6a118b37c522bbdf8a12e24e0486a234053db8" id=c34d6a47-595f-4cb5-a14e-b6448695ba61 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:27.600141946Z" level=info msg="runSandbox: unmounting shmPath for sandbox 6e0f82451fbe1ee8932e9ef9fb6a118b37c522bbdf8a12e24e0486a234053db8" id=c34d6a47-595f-4cb5-a14e-b6448695ba61 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:27.604025820Z" level=info msg="runSandbox: removing pod sandbox from storage: 6e0f82451fbe1ee8932e9ef9fb6a118b37c522bbdf8a12e24e0486a234053db8" id=c34d6a47-595f-4cb5-a14e-b6448695ba61 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:27.610486736Z" level=info msg="runSandbox: releasing container name: k8s_POD_cni-sysctl-allowlist-ds-ccgld_openshift-multus_42d075ab-9d7c-4a5d-8c34-2d2b27434266_0" id=c34d6a47-595f-4cb5-a14e-b6448695ba61 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:27.610512903Z" level=info msg="runSandbox: releasing pod sandbox name: k8s_cni-sysctl-allowlist-ds-ccgld_openshift-multus_42d075ab-9d7c-4a5d-8c34-2d2b27434266_0" id=c34d6a47-595f-4cb5-a14e-b6448695ba61 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:27.610739 2927 remote_runtime.go:222] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_cni-sysctl-allowlist-ds-ccgld_openshift-multus_42d075ab-9d7c-4a5d-8c34-2d2b27434266_0(6e0f82451fbe1ee8932e9ef9fb6a118b37c522bbdf8a12e24e0486a234053db8): error adding pod openshift-multus_cni-sysctl-allowlist-ds-ccgld to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-multus/cni-sysctl-allowlist-ds-ccgld/42d075ab-9d7c-4a5d-8c34-2d2b27434266]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:27.610811 2927 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_cni-sysctl-allowlist-ds-ccgld_openshift-multus_42d075ab-9d7c-4a5d-8c34-2d2b27434266_0(6e0f82451fbe1ee8932e9ef9fb6a118b37c522bbdf8a12e24e0486a234053db8): error adding pod openshift-multus_cni-sysctl-allowlist-ds-ccgld to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-multus/cni-sysctl-allowlist-ds-ccgld/42d075ab-9d7c-4a5d-8c34-2d2b27434266]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-multus/cni-sysctl-allowlist-ds-ccgld" Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:27.610872 2927 kuberuntime_manager.go:772] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_cni-sysctl-allowlist-ds-ccgld_openshift-multus_42d075ab-9d7c-4a5d-8c34-2d2b27434266_0(6e0f82451fbe1ee8932e9ef9fb6a118b37c522bbdf8a12e24e0486a234053db8): error adding pod openshift-multus_cni-sysctl-allowlist-ds-ccgld to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-multus/cni-sysctl-allowlist-ds-ccgld/42d075ab-9d7c-4a5d-8c34-2d2b27434266]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-multus/cni-sysctl-allowlist-ds-ccgld" Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:27.610976 2927 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"cni-sysctl-allowlist-ds-ccgld_openshift-multus(42d075ab-9d7c-4a5d-8c34-2d2b27434266)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"cni-sysctl-allowlist-ds-ccgld_openshift-multus(42d075ab-9d7c-4a5d-8c34-2d2b27434266)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_cni-sysctl-allowlist-ds-ccgld_openshift-multus_42d075ab-9d7c-4a5d-8c34-2d2b27434266_0(6e0f82451fbe1ee8932e9ef9fb6a118b37c522bbdf8a12e24e0486a234053db8): error adding pod openshift-multus_cni-sysctl-allowlist-ds-ccgld to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus\\\" name=\\\"multus-cni-network\\\" failed (add): Multus: [openshift-multus/cni-sysctl-allowlist-ds-ccgld/42d075ab-9d7c-4a5d-8c34-2d2b27434266]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition\"" pod="openshift-multus/cni-sysctl-allowlist-ds-ccgld" podUID=42d075ab-9d7c-4a5d-8c34-2d2b27434266 Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:27.654939 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-ccgld" Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:27.655338260Z" level=info msg="Running pod sandbox: openshift-multus/cni-sysctl-allowlist-ds-ccgld/POD" id=38fc5ec6-439a-4806-b223-6208b1edd1c9 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:27.655386062Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:27.673150260Z" level=info msg="Got pod network &{Name:cni-sysctl-allowlist-ds-ccgld Namespace:openshift-multus ID:b0357b2427dbf428186e352b2d924d648a9f435c4ee4f03812d158cd89c3b229 UID:42d075ab-9d7c-4a5d-8c34-2d2b27434266 NetNS:/var/run/netns/a211c907-eb76-4818-b869-f0533d0c01e0 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:27.673184997Z" level=info msg="Adding pod openshift-multus_cni-sysctl-allowlist-ds-ccgld to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:27.761954 2927 plugin_watcher.go:203] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/diskplugin.csi.alibabacloud.com-reg.sock" Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:27.950538 2927 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-multus/cni-sysctl-allowlist-ds-ccgld] Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: I0105 09:29:27.828830 6048 ovs.go:90] Maximum command line arguments set to: 191102 Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: 2023-01-05T09:29:27Z [verbose] Add: openshift-multus:cni-sysctl-allowlist-ds-ccgld:42d075ab-9d7c-4a5d-8c34-2d2b27434266:ovn-kubernetes(ovn-kubernetes):eth0 {"cniVersion":"0.4.0","interfaces":[{"name":"b0357b2427dbf42","mac":"da:b8:22:72:7b:d1"},{"name":"eth0","mac":"0a:58:0a:83:00:07","sandbox":"/var/run/netns/a211c907-eb76-4818-b869-f0533d0c01e0"}],"ips":[{"version":"4","interface":1,"address":"10.131.0.7/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: I0105 09:29:27.935502 6042 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-multus", Name:"cni-sysctl-allowlist-ds-ccgld", UID:"42d075ab-9d7c-4a5d-8c34-2d2b27434266", APIVersion:"v1", ResourceVersion:"90004", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.7/23] from ovn-kubernetes Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:27.951368444Z" level=info msg="Got pod network &{Name:cni-sysctl-allowlist-ds-ccgld Namespace:openshift-multus ID:b0357b2427dbf428186e352b2d924d648a9f435c4ee4f03812d158cd89c3b229 UID:42d075ab-9d7c-4a5d-8c34-2d2b27434266 NetNS:/var/run/netns/a211c907-eb76-4818-b869-f0533d0c01e0 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:27.951527240Z" level=info msg="Checking pod openshift-multus_cni-sysctl-allowlist-ds-ccgld for CNI network multus-cni-network (type=multus)" Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:29:27.954181 2927 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42d075ab_9d7c_4a5d_8c34_2d2b27434266.slice/crio-b0357b2427dbf428186e352b2d924d648a9f435c4ee4f03812d158cd89c3b229.scope WatchSource:0}: Error finding container b0357b2427dbf428186e352b2d924d648a9f435c4ee4f03812d158cd89c3b229: Status 404 returned error can't find the container with id b0357b2427dbf428186e352b2d924d648a9f435c4ee4f03812d158cd89c3b229 Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:27.957599007Z" level=info msg="Ran pod sandbox b0357b2427dbf428186e352b2d924d648a9f435c4ee4f03812d158cd89c3b229 with infra container: openshift-multus/cni-sysctl-allowlist-ds-ccgld/POD" id=38fc5ec6-439a-4806-b223-6208b1edd1c9 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:27.958426727Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=6b753d95-f3e6-4bba-9484-92aaa4084b27 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:27.958615154Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=6b753d95-f3e6-4bba-9484-92aaa4084b27 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:27.959307352Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=9dd53736-3764-43f4-8009-af22dc8a28dc name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:27.959461359Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=9dd53736-3764-43f4-8009-af22dc8a28dc name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:27.960200330Z" level=info msg="Creating container: openshift-multus/cni-sysctl-allowlist-ds-ccgld/kube-multus-additional-cni-plugins" id=864edf3f-417a-4548-ab2d-1dfb535409e0 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:27.960309441Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:28.111625 2927 reconciler.go:164] "OperationExecutor.RegisterPlugin started" plugin={SocketPath:/var/lib/kubelet/plugins_registry/diskplugin.csi.alibabacloud.com-reg.sock Timestamp:2023-01-05 09:29:27.761976462 +0000 UTC m=+299.517659375 Handler: Name:} Jan 05 09:29:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:28.112470755Z" level=info msg="Created container d4aa35750bb9cbb332c04ef2b3b110c8ccb79d5cb45a46daec15fffd82ab4b90: openshift-multus/cni-sysctl-allowlist-ds-ccgld/kube-multus-additional-cni-plugins" id=864edf3f-417a-4548-ab2d-1dfb535409e0 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:28.113201865Z" level=info msg="Starting container: d4aa35750bb9cbb332c04ef2b3b110c8ccb79d5cb45a46daec15fffd82ab4b90" id=9f845293-3fad-47d8-9e19-1c0e45daee75 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:28.114451 2927 csi_plugin.go:99] kubernetes.io/csi: Trying to validate a new CSI Driver with name: diskplugin.csi.alibabacloud.com endpoint: /var/lib/kubelet/plugins/diskplugin.csi.alibabacloud.com/csi.sock versions: 1.0.0 Jan 05 09:29:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:28.114484 2927 csi_plugin.go:112] kubernetes.io/csi: Register new plugin with name: diskplugin.csi.alibabacloud.com at endpoint: /var/lib/kubelet/plugins/diskplugin.csi.alibabacloud.com/csi.sock Jan 05 09:29:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:28.123803555Z" level=info msg="Started container" PID=6083 containerID=d4aa35750bb9cbb332c04ef2b3b110c8ccb79d5cb45a46daec15fffd82ab4b90 description=openshift-multus/cni-sysctl-allowlist-ds-ccgld/kube-multus-additional-cni-plugins id=9f845293-3fad-47d8-9e19-1c0e45daee75 name=/runtime.v1.RuntimeService/StartContainer sandboxID=b0357b2427dbf428186e352b2d924d648a9f435c4ee4f03812d158cd89c3b229 Jan 05 09:29:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:28.599336 2927 kubelet.go:1343] "Image garbage collection succeeded" Jan 05 09:29:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:28.657466 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-ccgld" event=&{ID:42d075ab-9d7c-4a5d-8c34-2d2b27434266 Type:ContainerStarted Data:d4aa35750bb9cbb332c04ef2b3b110c8ccb79d5cb45a46daec15fffd82ab4b90} Jan 05 09:29:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:28.657501 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-ccgld" event=&{ID:42d075ab-9d7c-4a5d-8c34-2d2b27434266 Type:ContainerStarted Data:b0357b2427dbf428186e352b2d924d648a9f435c4ee4f03812d158cd89c3b229} Jan 05 09:29:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:28.658026 2927 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-multus/cni-sysctl-allowlist-ds-ccgld" Jan 05 09:29:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:28.797229465Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=aae6d0c2-27ca-43e4-ab3d-7413800dddef name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:28.797479806Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=aae6d0c2-27ca-43e4-ab3d-7413800dddef name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:29.690508 2927 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-multus/cni-sysctl-allowlist-ds-ccgld" Jan 05 09:29:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:32.072834 2927 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-multus/cni-sysctl-allowlist-ds-ccgld] Jan 05 09:29:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:32.073032 2927 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-multus/cni-sysctl-allowlist-ds-ccgld" podUID=42d075ab-9d7c-4a5d-8c34-2d2b27434266 containerName="kube-multus-additional-cni-plugins" containerID="cri-o://d4aa35750bb9cbb332c04ef2b3b110c8ccb79d5cb45a46daec15fffd82ab4b90" gracePeriod=30 Jan 05 09:29:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:32.073606298Z" level=info msg="Stopping container: d4aa35750bb9cbb332c04ef2b3b110c8ccb79d5cb45a46daec15fffd82ab4b90 (timeout: 30s)" id=59a32b47-1fb0-4108-8b67-1fbf4af6457c name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.295682209Z" level=error msg="Error stopping network on cleanup: failed to destroy network for pod sandbox k8s_prometheus-adapter-5894cc86c7-qhmbh_openshift-monitoring_a40bf20d-67a4-45e7-b701-ca5a75376c1f_0(74f272f4c50be5a640d85c29438b61a9a814b4b072afe86593507911550e6a03): error removing pod openshift-monitoring_prometheus-adapter-5894cc86c7-qhmbh from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh/a40bf20d-67a4-45e7-b701-ca5a75376c1f]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" id=84bf72ae-8e03-4616-a642-c3d68a94a275 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.295729437Z" level=info msg="runSandbox: cleaning up namespaces after failing to run sandbox 74f272f4c50be5a640d85c29438b61a9a814b4b072afe86593507911550e6a03" id=84bf72ae-8e03-4616-a642-c3d68a94a275 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.310792478Z" level=error msg="Error stopping network on cleanup: failed to destroy network for pod sandbox k8s_prometheus-operator-admission-webhook-598468bb49-94dsn_openshift-monitoring_88745ab7-efe7-42c0-a2e8-75591518333d_0(ff598f9236d4f32150ffef17742e5f55d19d3b2adb4f7800e050a1ea04825617): error removing pod openshift-monitoring_prometheus-operator-admission-webhook-598468bb49-94dsn from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn/88745ab7-efe7-42c0-a2e8-75591518333d]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" id=b5f01488-c439-4bce-8eba-9abb8505890d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.310832623Z" level=info msg="runSandbox: cleaning up namespaces after failing to run sandbox ff598f9236d4f32150ffef17742e5f55d19d3b2adb4f7800e050a1ea04825617" id=b5f01488-c439-4bce-8eba-9abb8505890d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.331438374Z" level=info msg="runSandbox: deleting pod ID 74f272f4c50be5a640d85c29438b61a9a814b4b072afe86593507911550e6a03 from idIndex" id=84bf72ae-8e03-4616-a642-c3d68a94a275 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.331489780Z" level=info msg="runSandbox: removing pod sandbox 74f272f4c50be5a640d85c29438b61a9a814b4b072afe86593507911550e6a03" id=84bf72ae-8e03-4616-a642-c3d68a94a275 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.331517857Z" level=info msg="runSandbox: deleting container ID from idIndex for sandbox 74f272f4c50be5a640d85c29438b61a9a814b4b072afe86593507911550e6a03" id=84bf72ae-8e03-4616-a642-c3d68a94a275 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.331538186Z" level=info msg="runSandbox: unmounting shmPath for sandbox 74f272f4c50be5a640d85c29438b61a9a814b4b072afe86593507911550e6a03" id=84bf72ae-8e03-4616-a642-c3d68a94a275 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.335002476Z" level=info msg="runSandbox: removing pod sandbox from storage: 74f272f4c50be5a640d85c29438b61a9a814b4b072afe86593507911550e6a03" id=84bf72ae-8e03-4616-a642-c3d68a94a275 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.345954207Z" level=info msg="runSandbox: deleting pod ID ff598f9236d4f32150ffef17742e5f55d19d3b2adb4f7800e050a1ea04825617 from idIndex" id=b5f01488-c439-4bce-8eba-9abb8505890d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.345991282Z" level=info msg="runSandbox: removing pod sandbox ff598f9236d4f32150ffef17742e5f55d19d3b2adb4f7800e050a1ea04825617" id=b5f01488-c439-4bce-8eba-9abb8505890d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.346020912Z" level=info msg="runSandbox: deleting container ID from idIndex for sandbox ff598f9236d4f32150ffef17742e5f55d19d3b2adb4f7800e050a1ea04825617" id=b5f01488-c439-4bce-8eba-9abb8505890d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.346034731Z" level=info msg="runSandbox: unmounting shmPath for sandbox ff598f9236d4f32150ffef17742e5f55d19d3b2adb4f7800e050a1ea04825617" id=b5f01488-c439-4bce-8eba-9abb8505890d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.348945852Z" level=info msg="runSandbox: removing pod sandbox from storage: ff598f9236d4f32150ffef17742e5f55d19d3b2adb4f7800e050a1ea04825617" id=b5f01488-c439-4bce-8eba-9abb8505890d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.349955642Z" level=error msg="Error stopping network on cleanup: failed to destroy network for pod sandbox k8s_network-metrics-daemon-hq594_openshift-multus_50152296-9489-4fa8-aa42-7858debe1c08_0(1150fb360a7f91f41f9b7eb9840a298cfab29d821843d0d60688bc2927bdfece): error removing pod openshift-multus_network-metrics-daemon-hq594 from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-multus/network-metrics-daemon-hq594/50152296-9489-4fa8-aa42-7858debe1c08]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" id=3ef0c17d-bba5-488b-ad96-8e18ae386263 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.349989410Z" level=info msg="runSandbox: cleaning up namespaces after failing to run sandbox 1150fb360a7f91f41f9b7eb9840a298cfab29d821843d0d60688bc2927bdfece" id=3ef0c17d-bba5-488b-ad96-8e18ae386263 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.351044692Z" level=info msg="runSandbox: releasing container name: k8s_POD_prometheus-adapter-5894cc86c7-qhmbh_openshift-monitoring_a40bf20d-67a4-45e7-b701-ca5a75376c1f_0" id=84bf72ae-8e03-4616-a642-c3d68a94a275 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.351077775Z" level=info msg="runSandbox: releasing pod sandbox name: k8s_prometheus-adapter-5894cc86c7-qhmbh_openshift-monitoring_a40bf20d-67a4-45e7-b701-ca5a75376c1f_0" id=84bf72ae-8e03-4616-a642-c3d68a94a275 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.351304 2927 remote_runtime.go:222] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_prometheus-adapter-5894cc86c7-qhmbh_openshift-monitoring_a40bf20d-67a4-45e7-b701-ca5a75376c1f_0(74f272f4c50be5a640d85c29438b61a9a814b4b072afe86593507911550e6a03): error adding pod openshift-monitoring_prometheus-adapter-5894cc86c7-qhmbh to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh/a40bf20d-67a4-45e7-b701-ca5a75376c1f]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.351371 2927 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_prometheus-adapter-5894cc86c7-qhmbh_openshift-monitoring_a40bf20d-67a4-45e7-b701-ca5a75376c1f_0(74f272f4c50be5a640d85c29438b61a9a814b4b072afe86593507911550e6a03): error adding pod openshift-monitoring_prometheus-adapter-5894cc86c7-qhmbh to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh/a40bf20d-67a4-45e7-b701-ca5a75376c1f]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.351412 2927 kuberuntime_manager.go:772] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_prometheus-adapter-5894cc86c7-qhmbh_openshift-monitoring_a40bf20d-67a4-45e7-b701-ca5a75376c1f_0(74f272f4c50be5a640d85c29438b61a9a814b4b072afe86593507911550e6a03): error adding pod openshift-monitoring_prometheus-adapter-5894cc86c7-qhmbh to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh/a40bf20d-67a4-45e7-b701-ca5a75376c1f]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.351489 2927 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"prometheus-adapter-5894cc86c7-qhmbh_openshift-monitoring(a40bf20d-67a4-45e7-b701-ca5a75376c1f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"prometheus-adapter-5894cc86c7-qhmbh_openshift-monitoring(a40bf20d-67a4-45e7-b701-ca5a75376c1f)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_prometheus-adapter-5894cc86c7-qhmbh_openshift-monitoring_a40bf20d-67a4-45e7-b701-ca5a75376c1f_0(74f272f4c50be5a640d85c29438b61a9a814b4b072afe86593507911550e6a03): error adding pod openshift-monitoring_prometheus-adapter-5894cc86c7-qhmbh to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus\\\" name=\\\"multus-cni-network\\\" failed (add): Multus: [openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh/a40bf20d-67a4-45e7-b701-ca5a75376c1f]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition\"" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" podUID=a40bf20d-67a4-45e7-b701-ca5a75376c1f Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.376840701Z" level=info msg="runSandbox: releasing container name: k8s_POD_prometheus-operator-admission-webhook-598468bb49-94dsn_openshift-monitoring_88745ab7-efe7-42c0-a2e8-75591518333d_0" id=b5f01488-c439-4bce-8eba-9abb8505890d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.376892377Z" level=info msg="runSandbox: releasing pod sandbox name: k8s_prometheus-operator-admission-webhook-598468bb49-94dsn_openshift-monitoring_88745ab7-efe7-42c0-a2e8-75591518333d_0" id=b5f01488-c439-4bce-8eba-9abb8505890d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.377249 2927 remote_runtime.go:222] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_prometheus-operator-admission-webhook-598468bb49-94dsn_openshift-monitoring_88745ab7-efe7-42c0-a2e8-75591518333d_0(ff598f9236d4f32150ffef17742e5f55d19d3b2adb4f7800e050a1ea04825617): error adding pod openshift-monitoring_prometheus-operator-admission-webhook-598468bb49-94dsn to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn/88745ab7-efe7-42c0-a2e8-75591518333d]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.377314 2927 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_prometheus-operator-admission-webhook-598468bb49-94dsn_openshift-monitoring_88745ab7-efe7-42c0-a2e8-75591518333d_0(ff598f9236d4f32150ffef17742e5f55d19d3b2adb4f7800e050a1ea04825617): error adding pod openshift-monitoring_prometheus-operator-admission-webhook-598468bb49-94dsn to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn/88745ab7-efe7-42c0-a2e8-75591518333d]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.377358 2927 kuberuntime_manager.go:772] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_prometheus-operator-admission-webhook-598468bb49-94dsn_openshift-monitoring_88745ab7-efe7-42c0-a2e8-75591518333d_0(ff598f9236d4f32150ffef17742e5f55d19d3b2adb4f7800e050a1ea04825617): error adding pod openshift-monitoring_prometheus-operator-admission-webhook-598468bb49-94dsn to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn/88745ab7-efe7-42c0-a2e8-75591518333d]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.377436 2927 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"prometheus-operator-admission-webhook-598468bb49-94dsn_openshift-monitoring(88745ab7-efe7-42c0-a2e8-75591518333d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"prometheus-operator-admission-webhook-598468bb49-94dsn_openshift-monitoring(88745ab7-efe7-42c0-a2e8-75591518333d)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_prometheus-operator-admission-webhook-598468bb49-94dsn_openshift-monitoring_88745ab7-efe7-42c0-a2e8-75591518333d_0(ff598f9236d4f32150ffef17742e5f55d19d3b2adb4f7800e050a1ea04825617): error adding pod openshift-monitoring_prometheus-operator-admission-webhook-598468bb49-94dsn to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus\\\" name=\\\"multus-cni-network\\\" failed (add): Multus: [openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn/88745ab7-efe7-42c0-a2e8-75591518333d]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition\"" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" podUID=88745ab7-efe7-42c0-a2e8-75591518333d Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.405947789Z" level=info msg="runSandbox: deleting pod ID 1150fb360a7f91f41f9b7eb9840a298cfab29d821843d0d60688bc2927bdfece from idIndex" id=3ef0c17d-bba5-488b-ad96-8e18ae386263 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.405978202Z" level=info msg="runSandbox: removing pod sandbox 1150fb360a7f91f41f9b7eb9840a298cfab29d821843d0d60688bc2927bdfece" id=3ef0c17d-bba5-488b-ad96-8e18ae386263 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.406006457Z" level=info msg="runSandbox: deleting container ID from idIndex for sandbox 1150fb360a7f91f41f9b7eb9840a298cfab29d821843d0d60688bc2927bdfece" id=3ef0c17d-bba5-488b-ad96-8e18ae386263 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.406026622Z" level=info msg="runSandbox: unmounting shmPath for sandbox 1150fb360a7f91f41f9b7eb9840a298cfab29d821843d0d60688bc2927bdfece" id=3ef0c17d-bba5-488b-ad96-8e18ae386263 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.408341680Z" level=error msg="Error stopping network on cleanup: failed to destroy network for pod sandbox k8s_router-default-5cf6bbdff9-tnffs_openshift-ingress_c3b10ff4-a243-4a4e-b27d-f613e258666d_0(8a255d5b5b10cf99b23616f05d7fb3c48ec24229e1f6cc5ce6484e928e0a86af): error removing pod openshift-ingress_router-default-5cf6bbdff9-tnffs from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-ingress/router-default-5cf6bbdff9-tnffs/c3b10ff4-a243-4a4e-b27d-f613e258666d]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" id=cea2f99d-429b-485e-9f18-1ae520031c8a name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.408379972Z" level=info msg="runSandbox: cleaning up namespaces after failing to run sandbox 8a255d5b5b10cf99b23616f05d7fb3c48ec24229e1f6cc5ce6484e928e0a86af" id=cea2f99d-429b-485e-9f18-1ae520031c8a name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.409928927Z" level=info msg="runSandbox: removing pod sandbox from storage: 1150fb360a7f91f41f9b7eb9840a298cfab29d821843d0d60688bc2927bdfece" id=3ef0c17d-bba5-488b-ad96-8e18ae386263 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.427871852Z" level=info msg="runSandbox: releasing container name: k8s_POD_network-metrics-daemon-hq594_openshift-multus_50152296-9489-4fa8-aa42-7858debe1c08_0" id=3ef0c17d-bba5-488b-ad96-8e18ae386263 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.427934476Z" level=info msg="runSandbox: releasing pod sandbox name: k8s_network-metrics-daemon-hq594_openshift-multus_50152296-9489-4fa8-aa42-7858debe1c08_0" id=3ef0c17d-bba5-488b-ad96-8e18ae386263 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.429637889Z" level=error msg="Error stopping network on cleanup: failed to destroy network for pod sandbox k8s_prometheus-k8s-0_openshift-monitoring_4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879_0(52b90175ef0a6f25e9a947f029d3b67d80d2c31734082f3c71e38fa2ba264d89): error removing pod openshift-monitoring_prometheus-k8s-0 from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-monitoring/prometheus-k8s-0/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" id=0e4a725d-2cd9-4f62-8e17-541b1dbb4f78 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.429681138Z" level=info msg="runSandbox: cleaning up namespaces after failing to run sandbox 52b90175ef0a6f25e9a947f029d3b67d80d2c31734082f3c71e38fa2ba264d89" id=0e4a725d-2cd9-4f62-8e17-541b1dbb4f78 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.430916 2927 remote_runtime.go:222] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_network-metrics-daemon-hq594_openshift-multus_50152296-9489-4fa8-aa42-7858debe1c08_0(1150fb360a7f91f41f9b7eb9840a298cfab29d821843d0d60688bc2927bdfece): error adding pod openshift-multus_network-metrics-daemon-hq594 to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-multus/network-metrics-daemon-hq594/50152296-9489-4fa8-aa42-7858debe1c08]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.430978 2927 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_network-metrics-daemon-hq594_openshift-multus_50152296-9489-4fa8-aa42-7858debe1c08_0(1150fb360a7f91f41f9b7eb9840a298cfab29d821843d0d60688bc2927bdfece): error adding pod openshift-multus_network-metrics-daemon-hq594 to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-multus/network-metrics-daemon-hq594/50152296-9489-4fa8-aa42-7858debe1c08]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.431016 2927 kuberuntime_manager.go:772] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_network-metrics-daemon-hq594_openshift-multus_50152296-9489-4fa8-aa42-7858debe1c08_0(1150fb360a7f91f41f9b7eb9840a298cfab29d821843d0d60688bc2927bdfece): error adding pod openshift-multus_network-metrics-daemon-hq594 to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-multus/network-metrics-daemon-hq594/50152296-9489-4fa8-aa42-7858debe1c08]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.431099 2927 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"network-metrics-daemon-hq594_openshift-multus(50152296-9489-4fa8-aa42-7858debe1c08)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"network-metrics-daemon-hq594_openshift-multus(50152296-9489-4fa8-aa42-7858debe1c08)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_network-metrics-daemon-hq594_openshift-multus_50152296-9489-4fa8-aa42-7858debe1c08_0(1150fb360a7f91f41f9b7eb9840a298cfab29d821843d0d60688bc2927bdfece): error adding pod openshift-multus_network-metrics-daemon-hq594 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus\\\" name=\\\"multus-cni-network\\\" failed (add): Multus: [openshift-multus/network-metrics-daemon-hq594/50152296-9489-4fa8-aa42-7858debe1c08]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition\"" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.441726612Z" level=error msg="Error stopping network on cleanup: failed to destroy network for pod sandbox k8s_alertmanager-main-0_openshift-monitoring_05c9a6ea-140a-49a2-bde5-f220b05aa252_0(ed0ee704407b6914733819122eb9fc60ca3996f0c1b935170d5903732b869347): error removing pod openshift-monitoring_alertmanager-main-0 from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-monitoring/alertmanager-main-0/05c9a6ea-140a-49a2-bde5-f220b05aa252]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" id=bc650832-6cc0-4148-bdb8-05b5ae4deaf1 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.441772098Z" level=info msg="runSandbox: cleaning up namespaces after failing to run sandbox ed0ee704407b6914733819122eb9fc60ca3996f0c1b935170d5903732b869347" id=bc650832-6cc0-4148-bdb8-05b5ae4deaf1 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.446945211Z" level=info msg="runSandbox: deleting pod ID 8a255d5b5b10cf99b23616f05d7fb3c48ec24229e1f6cc5ce6484e928e0a86af from idIndex" id=cea2f99d-429b-485e-9f18-1ae520031c8a name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.446985717Z" level=info msg="runSandbox: removing pod sandbox 8a255d5b5b10cf99b23616f05d7fb3c48ec24229e1f6cc5ce6484e928e0a86af" id=cea2f99d-429b-485e-9f18-1ae520031c8a name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.447009214Z" level=info msg="runSandbox: deleting container ID from idIndex for sandbox 8a255d5b5b10cf99b23616f05d7fb3c48ec24229e1f6cc5ce6484e928e0a86af" id=cea2f99d-429b-485e-9f18-1ae520031c8a name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.447029575Z" level=info msg="runSandbox: unmounting shmPath for sandbox 8a255d5b5b10cf99b23616f05d7fb3c48ec24229e1f6cc5ce6484e928e0a86af" id=cea2f99d-429b-485e-9f18-1ae520031c8a name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.449928069Z" level=info msg="runSandbox: removing pod sandbox from storage: 8a255d5b5b10cf99b23616f05d7fb3c48ec24229e1f6cc5ce6484e928e0a86af" id=cea2f99d-429b-485e-9f18-1ae520031c8a name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.456202802Z" level=info msg="runSandbox: releasing container name: k8s_POD_router-default-5cf6bbdff9-tnffs_openshift-ingress_c3b10ff4-a243-4a4e-b27d-f613e258666d_0" id=cea2f99d-429b-485e-9f18-1ae520031c8a name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.456227420Z" level=info msg="runSandbox: releasing pod sandbox name: k8s_router-default-5cf6bbdff9-tnffs_openshift-ingress_c3b10ff4-a243-4a4e-b27d-f613e258666d_0" id=cea2f99d-429b-485e-9f18-1ae520031c8a name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.456418 2927 remote_runtime.go:222] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_router-default-5cf6bbdff9-tnffs_openshift-ingress_c3b10ff4-a243-4a4e-b27d-f613e258666d_0(8a255d5b5b10cf99b23616f05d7fb3c48ec24229e1f6cc5ce6484e928e0a86af): error adding pod openshift-ingress_router-default-5cf6bbdff9-tnffs to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-ingress/router-default-5cf6bbdff9-tnffs/c3b10ff4-a243-4a4e-b27d-f613e258666d]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.456483 2927 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_router-default-5cf6bbdff9-tnffs_openshift-ingress_c3b10ff4-a243-4a4e-b27d-f613e258666d_0(8a255d5b5b10cf99b23616f05d7fb3c48ec24229e1f6cc5ce6484e928e0a86af): error adding pod openshift-ingress_router-default-5cf6bbdff9-tnffs to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-ingress/router-default-5cf6bbdff9-tnffs/c3b10ff4-a243-4a4e-b27d-f613e258666d]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.456518 2927 kuberuntime_manager.go:772] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_router-default-5cf6bbdff9-tnffs_openshift-ingress_c3b10ff4-a243-4a4e-b27d-f613e258666d_0(8a255d5b5b10cf99b23616f05d7fb3c48ec24229e1f6cc5ce6484e928e0a86af): error adding pod openshift-ingress_router-default-5cf6bbdff9-tnffs to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-ingress/router-default-5cf6bbdff9-tnffs/c3b10ff4-a243-4a4e-b27d-f613e258666d]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.456597 2927 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"router-default-5cf6bbdff9-tnffs_openshift-ingress(c3b10ff4-a243-4a4e-b27d-f613e258666d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"router-default-5cf6bbdff9-tnffs_openshift-ingress(c3b10ff4-a243-4a4e-b27d-f613e258666d)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_router-default-5cf6bbdff9-tnffs_openshift-ingress_c3b10ff4-a243-4a4e-b27d-f613e258666d_0(8a255d5b5b10cf99b23616f05d7fb3c48ec24229e1f6cc5ce6484e928e0a86af): error adding pod openshift-ingress_router-default-5cf6bbdff9-tnffs to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus\\\" name=\\\"multus-cni-network\\\" failed (add): Multus: [openshift-ingress/router-default-5cf6bbdff9-tnffs/c3b10ff4-a243-4a4e-b27d-f613e258666d]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition\"" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" podUID=c3b10ff4-a243-4a4e-b27d-f613e258666d Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.461955755Z" level=info msg="runSandbox: deleting pod ID 52b90175ef0a6f25e9a947f029d3b67d80d2c31734082f3c71e38fa2ba264d89 from idIndex" id=0e4a725d-2cd9-4f62-8e17-541b1dbb4f78 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.461983810Z" level=info msg="runSandbox: removing pod sandbox 52b90175ef0a6f25e9a947f029d3b67d80d2c31734082f3c71e38fa2ba264d89" id=0e4a725d-2cd9-4f62-8e17-541b1dbb4f78 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.462003790Z" level=info msg="runSandbox: deleting container ID from idIndex for sandbox 52b90175ef0a6f25e9a947f029d3b67d80d2c31734082f3c71e38fa2ba264d89" id=0e4a725d-2cd9-4f62-8e17-541b1dbb4f78 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.462024908Z" level=info msg="runSandbox: unmounting shmPath for sandbox 52b90175ef0a6f25e9a947f029d3b67d80d2c31734082f3c71e38fa2ba264d89" id=0e4a725d-2cd9-4f62-8e17-541b1dbb4f78 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.464935696Z" level=info msg="runSandbox: removing pod sandbox from storage: 52b90175ef0a6f25e9a947f029d3b67d80d2c31734082f3c71e38fa2ba264d89" id=0e4a725d-2cd9-4f62-8e17-541b1dbb4f78 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.483381340Z" level=info msg="runSandbox: releasing container name: k8s_POD_prometheus-k8s-0_openshift-monitoring_4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879_0" id=0e4a725d-2cd9-4f62-8e17-541b1dbb4f78 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.483405265Z" level=info msg="runSandbox: releasing pod sandbox name: k8s_prometheus-k8s-0_openshift-monitoring_4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879_0" id=0e4a725d-2cd9-4f62-8e17-541b1dbb4f78 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.483616 2927 remote_runtime.go:222] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_prometheus-k8s-0_openshift-monitoring_4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879_0(52b90175ef0a6f25e9a947f029d3b67d80d2c31734082f3c71e38fa2ba264d89): error adding pod openshift-monitoring_prometheus-k8s-0 to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/prometheus-k8s-0/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.483673 2927 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_prometheus-k8s-0_openshift-monitoring_4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879_0(52b90175ef0a6f25e9a947f029d3b67d80d2c31734082f3c71e38fa2ba264d89): error adding pod openshift-monitoring_prometheus-k8s-0 to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/prometheus-k8s-0/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.483709 2927 kuberuntime_manager.go:772] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_prometheus-k8s-0_openshift-monitoring_4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879_0(52b90175ef0a6f25e9a947f029d3b67d80d2c31734082f3c71e38fa2ba264d89): error adding pod openshift-monitoring_prometheus-k8s-0 to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/prometheus-k8s-0/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.483783 2927 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"prometheus-k8s-0_openshift-monitoring(4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"prometheus-k8s-0_openshift-monitoring(4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_prometheus-k8s-0_openshift-monitoring_4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879_0(52b90175ef0a6f25e9a947f029d3b67d80d2c31734082f3c71e38fa2ba264d89): error adding pod openshift-monitoring_prometheus-k8s-0 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus\\\" name=\\\"multus-cni-network\\\" failed (add): Multus: [openshift-monitoring/prometheus-k8s-0/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition\"" pod="openshift-monitoring/prometheus-k8s-0" podUID=4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.489033836Z" level=info msg="runSandbox: deleting pod ID ed0ee704407b6914733819122eb9fc60ca3996f0c1b935170d5903732b869347 from idIndex" id=bc650832-6cc0-4148-bdb8-05b5ae4deaf1 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.489061703Z" level=info msg="runSandbox: removing pod sandbox ed0ee704407b6914733819122eb9fc60ca3996f0c1b935170d5903732b869347" id=bc650832-6cc0-4148-bdb8-05b5ae4deaf1 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.489082047Z" level=info msg="runSandbox: deleting container ID from idIndex for sandbox ed0ee704407b6914733819122eb9fc60ca3996f0c1b935170d5903732b869347" id=bc650832-6cc0-4148-bdb8-05b5ae4deaf1 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.489102470Z" level=info msg="runSandbox: unmounting shmPath for sandbox ed0ee704407b6914733819122eb9fc60ca3996f0c1b935170d5903732b869347" id=bc650832-6cc0-4148-bdb8-05b5ae4deaf1 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.492928286Z" level=info msg="runSandbox: removing pod sandbox from storage: ed0ee704407b6914733819122eb9fc60ca3996f0c1b935170d5903732b869347" id=bc650832-6cc0-4148-bdb8-05b5ae4deaf1 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.499040145Z" level=info msg="runSandbox: releasing container name: k8s_POD_alertmanager-main-0_openshift-monitoring_05c9a6ea-140a-49a2-bde5-f220b05aa252_0" id=bc650832-6cc0-4148-bdb8-05b5ae4deaf1 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.499065720Z" level=info msg="runSandbox: releasing pod sandbox name: k8s_alertmanager-main-0_openshift-monitoring_05c9a6ea-140a-49a2-bde5-f220b05aa252_0" id=bc650832-6cc0-4148-bdb8-05b5ae4deaf1 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.499234 2927 remote_runtime.go:222] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_alertmanager-main-0_openshift-monitoring_05c9a6ea-140a-49a2-bde5-f220b05aa252_0(ed0ee704407b6914733819122eb9fc60ca3996f0c1b935170d5903732b869347): error adding pod openshift-monitoring_alertmanager-main-0 to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/alertmanager-main-0/05c9a6ea-140a-49a2-bde5-f220b05aa252]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.499289 2927 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_alertmanager-main-0_openshift-monitoring_05c9a6ea-140a-49a2-bde5-f220b05aa252_0(ed0ee704407b6914733819122eb9fc60ca3996f0c1b935170d5903732b869347): error adding pod openshift-monitoring_alertmanager-main-0 to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/alertmanager-main-0/05c9a6ea-140a-49a2-bde5-f220b05aa252]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.499327 2927 kuberuntime_manager.go:772] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_alertmanager-main-0_openshift-monitoring_05c9a6ea-140a-49a2-bde5-f220b05aa252_0(ed0ee704407b6914733819122eb9fc60ca3996f0c1b935170d5903732b869347): error adding pod openshift-monitoring_alertmanager-main-0 to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/alertmanager-main-0/05c9a6ea-140a-49a2-bde5-f220b05aa252]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.499402 2927 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"alertmanager-main-0_openshift-monitoring(05c9a6ea-140a-49a2-bde5-f220b05aa252)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"alertmanager-main-0_openshift-monitoring(05c9a6ea-140a-49a2-bde5-f220b05aa252)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_alertmanager-main-0_openshift-monitoring_05c9a6ea-140a-49a2-bde5-f220b05aa252_0(ed0ee704407b6914733819122eb9fc60ca3996f0c1b935170d5903732b869347): error adding pod openshift-monitoring_alertmanager-main-0 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus\\\" name=\\\"multus-cni-network\\\" failed (add): Multus: [openshift-monitoring/alertmanager-main-0/05c9a6ea-140a-49a2-bde5-f220b05aa252]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition\"" pod="openshift-monitoring/alertmanager-main-0" podUID=05c9a6ea-140a-49a2-bde5-f220b05aa252 Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.535016681Z" level=error msg="Error stopping network on cleanup: failed to destroy network for pod sandbox k8s_ingress-canary-8xrbh_openshift-ingress-canary_e04a47b4-aa8e-4d4e-9d45-aa8037dcb748_0(2c2a604331152b239cd95403d2bb266e2ca816b6a0414973001dc448b2d2b416): error removing pod openshift-ingress-canary_ingress-canary-8xrbh from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-ingress-canary/ingress-canary-8xrbh/e04a47b4-aa8e-4d4e-9d45-aa8037dcb748]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" id=606b9627-db1b-4a3f-874b-b462a6d88dd8 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.535062283Z" level=info msg="runSandbox: cleaning up namespaces after failing to run sandbox 2c2a604331152b239cd95403d2bb266e2ca816b6a0414973001dc448b2d2b416" id=606b9627-db1b-4a3f-874b-b462a6d88dd8 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.582935070Z" level=info msg="runSandbox: deleting pod ID 2c2a604331152b239cd95403d2bb266e2ca816b6a0414973001dc448b2d2b416 from idIndex" id=606b9627-db1b-4a3f-874b-b462a6d88dd8 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.582973864Z" level=info msg="runSandbox: removing pod sandbox 2c2a604331152b239cd95403d2bb266e2ca816b6a0414973001dc448b2d2b416" id=606b9627-db1b-4a3f-874b-b462a6d88dd8 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.582995790Z" level=info msg="runSandbox: deleting container ID from idIndex for sandbox 2c2a604331152b239cd95403d2bb266e2ca816b6a0414973001dc448b2d2b416" id=606b9627-db1b-4a3f-874b-b462a6d88dd8 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.583014808Z" level=info msg="runSandbox: unmounting shmPath for sandbox 2c2a604331152b239cd95403d2bb266e2ca816b6a0414973001dc448b2d2b416" id=606b9627-db1b-4a3f-874b-b462a6d88dd8 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.587933295Z" level=info msg="runSandbox: removing pod sandbox from storage: 2c2a604331152b239cd95403d2bb266e2ca816b6a0414973001dc448b2d2b416" id=606b9627-db1b-4a3f-874b-b462a6d88dd8 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.608427050Z" level=error msg="Error stopping network on cleanup: failed to destroy network for pod sandbox k8s_telemeter-client-5598c645c4-mftdv_openshift-monitoring_fc45e847-8f55-4c48-8c48-4bfee2419dca_0(a16c8f8a13d2354862769ea877c06e032be034fdb676b065e5595fbceae680cf): error removing pod openshift-monitoring_telemeter-client-5598c645c4-mftdv from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-monitoring/telemeter-client-5598c645c4-mftdv/fc45e847-8f55-4c48-8c48-4bfee2419dca]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" id=64b71f4f-3284-4db9-a02a-4cd2ff5dd755 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.608477563Z" level=info msg="runSandbox: cleaning up namespaces after failing to run sandbox a16c8f8a13d2354862769ea877c06e032be034fdb676b065e5595fbceae680cf" id=64b71f4f-3284-4db9-a02a-4cd2ff5dd755 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.612284692Z" level=info msg="runSandbox: releasing container name: k8s_POD_ingress-canary-8xrbh_openshift-ingress-canary_e04a47b4-aa8e-4d4e-9d45-aa8037dcb748_0" id=606b9627-db1b-4a3f-874b-b462a6d88dd8 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.612325445Z" level=info msg="runSandbox: releasing pod sandbox name: k8s_ingress-canary-8xrbh_openshift-ingress-canary_e04a47b4-aa8e-4d4e-9d45-aa8037dcb748_0" id=606b9627-db1b-4a3f-874b-b462a6d88dd8 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.612753 2927 remote_runtime.go:222] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_ingress-canary-8xrbh_openshift-ingress-canary_e04a47b4-aa8e-4d4e-9d45-aa8037dcb748_0(2c2a604331152b239cd95403d2bb266e2ca816b6a0414973001dc448b2d2b416): error adding pod openshift-ingress-canary_ingress-canary-8xrbh to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-ingress-canary/ingress-canary-8xrbh/e04a47b4-aa8e-4d4e-9d45-aa8037dcb748]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.612827 2927 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_ingress-canary-8xrbh_openshift-ingress-canary_e04a47b4-aa8e-4d4e-9d45-aa8037dcb748_0(2c2a604331152b239cd95403d2bb266e2ca816b6a0414973001dc448b2d2b416): error adding pod openshift-ingress-canary_ingress-canary-8xrbh to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-ingress-canary/ingress-canary-8xrbh/e04a47b4-aa8e-4d4e-9d45-aa8037dcb748]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-ingress-canary/ingress-canary-8xrbh" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.612869 2927 kuberuntime_manager.go:772] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_ingress-canary-8xrbh_openshift-ingress-canary_e04a47b4-aa8e-4d4e-9d45-aa8037dcb748_0(2c2a604331152b239cd95403d2bb266e2ca816b6a0414973001dc448b2d2b416): error adding pod openshift-ingress-canary_ingress-canary-8xrbh to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-ingress-canary/ingress-canary-8xrbh/e04a47b4-aa8e-4d4e-9d45-aa8037dcb748]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-ingress-canary/ingress-canary-8xrbh" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.612986 2927 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"ingress-canary-8xrbh_openshift-ingress-canary(e04a47b4-aa8e-4d4e-9d45-aa8037dcb748)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"ingress-canary-8xrbh_openshift-ingress-canary(e04a47b4-aa8e-4d4e-9d45-aa8037dcb748)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_ingress-canary-8xrbh_openshift-ingress-canary_e04a47b4-aa8e-4d4e-9d45-aa8037dcb748_0(2c2a604331152b239cd95403d2bb266e2ca816b6a0414973001dc448b2d2b416): error adding pod openshift-ingress-canary_ingress-canary-8xrbh to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus\\\" name=\\\"multus-cni-network\\\" failed (add): Multus: [openshift-ingress-canary/ingress-canary-8xrbh/e04a47b4-aa8e-4d4e-9d45-aa8037dcb748]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition\"" pod="openshift-ingress-canary/ingress-canary-8xrbh" podUID=e04a47b4-aa8e-4d4e-9d45-aa8037dcb748 Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.619082745Z" level=error msg="Error stopping network on cleanup: failed to destroy network for pod sandbox k8s_dns-default-2pfzf_openshift-dns_1cc2aadb-14f2-4d6e-97fc-6fdab8889a63_0(dcf02c2c8b254c0ba075cbb4447c97111fedbd18176c9b10dcf0b8e0fa4c2b1d): error removing pod openshift-dns_dns-default-2pfzf from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-dns/dns-default-2pfzf/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" id=e1a3ca46-101b-43fa-9721-c16b2c482369 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.619125898Z" level=info msg="runSandbox: cleaning up namespaces after failing to run sandbox dcf02c2c8b254c0ba075cbb4447c97111fedbd18176c9b10dcf0b8e0fa4c2b1d" id=e1a3ca46-101b-43fa-9721-c16b2c482369 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.635947725Z" level=info msg="runSandbox: deleting pod ID a16c8f8a13d2354862769ea877c06e032be034fdb676b065e5595fbceae680cf from idIndex" id=64b71f4f-3284-4db9-a02a-4cd2ff5dd755 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.635984832Z" level=info msg="runSandbox: removing pod sandbox a16c8f8a13d2354862769ea877c06e032be034fdb676b065e5595fbceae680cf" id=64b71f4f-3284-4db9-a02a-4cd2ff5dd755 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.636008848Z" level=info msg="runSandbox: deleting container ID from idIndex for sandbox a16c8f8a13d2354862769ea877c06e032be034fdb676b065e5595fbceae680cf" id=64b71f4f-3284-4db9-a02a-4cd2ff5dd755 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.636029434Z" level=info msg="runSandbox: unmounting shmPath for sandbox a16c8f8a13d2354862769ea877c06e032be034fdb676b065e5595fbceae680cf" id=64b71f4f-3284-4db9-a02a-4cd2ff5dd755 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.652947605Z" level=info msg="runSandbox: removing pod sandbox from storage: a16c8f8a13d2354862769ea877c06e032be034fdb676b065e5595fbceae680cf" id=64b71f4f-3284-4db9-a02a-4cd2ff5dd755 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.659057143Z" level=info msg="runSandbox: releasing container name: k8s_POD_telemeter-client-5598c645c4-mftdv_openshift-monitoring_fc45e847-8f55-4c48-8c48-4bfee2419dca_0" id=64b71f4f-3284-4db9-a02a-4cd2ff5dd755 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.659083893Z" level=info msg="runSandbox: releasing pod sandbox name: k8s_telemeter-client-5598c645c4-mftdv_openshift-monitoring_fc45e847-8f55-4c48-8c48-4bfee2419dca_0" id=64b71f4f-3284-4db9-a02a-4cd2ff5dd755 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.659284 2927 remote_runtime.go:222] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_telemeter-client-5598c645c4-mftdv_openshift-monitoring_fc45e847-8f55-4c48-8c48-4bfee2419dca_0(a16c8f8a13d2354862769ea877c06e032be034fdb676b065e5595fbceae680cf): error adding pod openshift-monitoring_telemeter-client-5598c645c4-mftdv to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/telemeter-client-5598c645c4-mftdv/fc45e847-8f55-4c48-8c48-4bfee2419dca]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.659349 2927 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_telemeter-client-5598c645c4-mftdv_openshift-monitoring_fc45e847-8f55-4c48-8c48-4bfee2419dca_0(a16c8f8a13d2354862769ea877c06e032be034fdb676b065e5595fbceae680cf): error adding pod openshift-monitoring_telemeter-client-5598c645c4-mftdv to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/telemeter-client-5598c645c4-mftdv/fc45e847-8f55-4c48-8c48-4bfee2419dca]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.659388 2927 kuberuntime_manager.go:772] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_telemeter-client-5598c645c4-mftdv_openshift-monitoring_fc45e847-8f55-4c48-8c48-4bfee2419dca_0(a16c8f8a13d2354862769ea877c06e032be034fdb676b065e5595fbceae680cf): error adding pod openshift-monitoring_telemeter-client-5598c645c4-mftdv to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/telemeter-client-5598c645c4-mftdv/fc45e847-8f55-4c48-8c48-4bfee2419dca]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.659469 2927 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"telemeter-client-5598c645c4-mftdv_openshift-monitoring(fc45e847-8f55-4c48-8c48-4bfee2419dca)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"telemeter-client-5598c645c4-mftdv_openshift-monitoring(fc45e847-8f55-4c48-8c48-4bfee2419dca)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_telemeter-client-5598c645c4-mftdv_openshift-monitoring_fc45e847-8f55-4c48-8c48-4bfee2419dca_0(a16c8f8a13d2354862769ea877c06e032be034fdb676b065e5595fbceae680cf): error adding pod openshift-monitoring_telemeter-client-5598c645c4-mftdv to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus\\\" name=\\\"multus-cni-network\\\" failed (add): Multus: [openshift-monitoring/telemeter-client-5598c645c4-mftdv/fc45e847-8f55-4c48-8c48-4bfee2419dca]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition\"" pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" podUID=fc45e847-8f55-4c48-8c48-4bfee2419dca Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.660943345Z" level=info msg="runSandbox: deleting pod ID dcf02c2c8b254c0ba075cbb4447c97111fedbd18176c9b10dcf0b8e0fa4c2b1d from idIndex" id=e1a3ca46-101b-43fa-9721-c16b2c482369 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.660980679Z" level=info msg="runSandbox: removing pod sandbox dcf02c2c8b254c0ba075cbb4447c97111fedbd18176c9b10dcf0b8e0fa4c2b1d" id=e1a3ca46-101b-43fa-9721-c16b2c482369 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.661002279Z" level=info msg="runSandbox: deleting container ID from idIndex for sandbox dcf02c2c8b254c0ba075cbb4447c97111fedbd18176c9b10dcf0b8e0fa4c2b1d" id=e1a3ca46-101b-43fa-9721-c16b2c482369 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.661022940Z" level=info msg="runSandbox: unmounting shmPath for sandbox dcf02c2c8b254c0ba075cbb4447c97111fedbd18176c9b10dcf0b8e0fa4c2b1d" id=e1a3ca46-101b-43fa-9721-c16b2c482369 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.663941839Z" level=info msg="runSandbox: removing pod sandbox from storage: dcf02c2c8b254c0ba075cbb4447c97111fedbd18176c9b10dcf0b8e0fa4c2b1d" id=e1a3ca46-101b-43fa-9721-c16b2c482369 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:35.667713 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:35.668236 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-8xrbh" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:35.668677 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:35.669099 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:35.669445 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:35.669781 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:35.670156 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:35.670589 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.674626994Z" level=info msg="Running pod sandbox: openshift-monitoring/prometheus-k8s-0/POD" id=e2aefa06-9295-4ff6-9b6b-ff22dd9746c2 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.674681953Z" level=info msg="Running pod sandbox: openshift-monitoring/telemeter-client-5598c645c4-mftdv/POD" id=2240f64b-1380-422d-8c5e-4835cf6a133c name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.674718941Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.674740334Z" level=info msg="Running pod sandbox: openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh/POD" id=f8a4f035-df5c-4a67-95f7-6afbfb299b6e name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.674767811Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.674682330Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.674860245Z" level=info msg="Running pod sandbox: openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn/POD" id=f9385c94-5f07-474c-a4be-b65b96cb4672 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.674928063Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.675022696Z" level=info msg="Running pod sandbox: openshift-multus/network-metrics-daemon-hq594/POD" id=3735d57a-2221-4f15-af34-ce6f61dcab84 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.675055068Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.675138712Z" level=info msg="Running pod sandbox: openshift-monitoring/alertmanager-main-0/POD" id=40151f20-e5d6-484a-a00f-ed789a9bb157 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.675154430Z" level=info msg="Running pod sandbox: openshift-ingress/router-default-5cf6bbdff9-tnffs/POD" id=98a61853-3c1c-45d2-9d24-5548fa947bdc name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.675170200Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.675181328Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.675265380Z" level=info msg="Running pod sandbox: openshift-ingress-canary/ingress-canary-8xrbh/POD" id=900808c6-6154-478e-8e45-2a15867c06a1 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.675292602Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.681206856Z" level=info msg="runSandbox: releasing container name: k8s_POD_dns-default-2pfzf_openshift-dns_1cc2aadb-14f2-4d6e-97fc-6fdab8889a63_0" id=e1a3ca46-101b-43fa-9721-c16b2c482369 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.681240031Z" level=info msg="runSandbox: releasing pod sandbox name: k8s_dns-default-2pfzf_openshift-dns_1cc2aadb-14f2-4d6e-97fc-6fdab8889a63_0" id=e1a3ca46-101b-43fa-9721-c16b2c482369 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.682528 2927 remote_runtime.go:222] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_dns-default-2pfzf_openshift-dns_1cc2aadb-14f2-4d6e-97fc-6fdab8889a63_0(dcf02c2c8b254c0ba075cbb4447c97111fedbd18176c9b10dcf0b8e0fa4c2b1d): error adding pod openshift-dns_dns-default-2pfzf to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-dns/dns-default-2pfzf/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.682576 2927 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_dns-default-2pfzf_openshift-dns_1cc2aadb-14f2-4d6e-97fc-6fdab8889a63_0(dcf02c2c8b254c0ba075cbb4447c97111fedbd18176c9b10dcf0b8e0fa4c2b1d): error adding pod openshift-dns_dns-default-2pfzf to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-dns/dns-default-2pfzf/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-dns/dns-default-2pfzf" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.682613 2927 kuberuntime_manager.go:772] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_dns-default-2pfzf_openshift-dns_1cc2aadb-14f2-4d6e-97fc-6fdab8889a63_0(dcf02c2c8b254c0ba075cbb4447c97111fedbd18176c9b10dcf0b8e0fa4c2b1d): error adding pod openshift-dns_dns-default-2pfzf to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-dns/dns-default-2pfzf/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-dns/dns-default-2pfzf" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.682694 2927 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"dns-default-2pfzf_openshift-dns(1cc2aadb-14f2-4d6e-97fc-6fdab8889a63)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"dns-default-2pfzf_openshift-dns(1cc2aadb-14f2-4d6e-97fc-6fdab8889a63)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_dns-default-2pfzf_openshift-dns_1cc2aadb-14f2-4d6e-97fc-6fdab8889a63_0(dcf02c2c8b254c0ba075cbb4447c97111fedbd18176c9b10dcf0b8e0fa4c2b1d): error adding pod openshift-dns_dns-default-2pfzf to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus\\\" name=\\\"multus-cni-network\\\" failed (add): Multus: [openshift-dns/dns-default-2pfzf/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition\"" pod="openshift-dns/dns-default-2pfzf" podUID=1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.706997919Z" level=error msg="Error stopping network on cleanup: failed to destroy network for pod sandbox k8s_kube-state-metrics-6798b548b9-xl88q_openshift-monitoring_2ecd2e2c-ab52-42e0-8fde-64e45804cc9e_0(a937bd9597c90954ad3041a6e3670bd2dfc453dbbd7b4d1b6d48e35dae7b83c0): error removing pod openshift-monitoring_kube-state-metrics-6798b548b9-xl88q from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" id=0755a791-7501-4417-934f-53470d89bb87 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.707038361Z" level=info msg="runSandbox: cleaning up namespaces after failing to run sandbox a937bd9597c90954ad3041a6e3670bd2dfc453dbbd7b4d1b6d48e35dae7b83c0" id=0755a791-7501-4417-934f-53470d89bb87 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.743374424Z" level=info msg="Got pod network &{Name:telemeter-client-5598c645c4-mftdv Namespace:openshift-monitoring ID:f1e02d518b5e62edc187ed7cc730fab2279f2eb5384e3f3ffd25d3a014b31d4b UID:fc45e847-8f55-4c48-8c48-4bfee2419dca NetNS:/var/run/netns/04b13e36-0acc-4f75-9752-0ebb73acd4ae Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.743409862Z" level=info msg="Adding pod openshift-monitoring_telemeter-client-5598c645c4-mftdv to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.743491592Z" level=info msg="Got pod network &{Name:network-metrics-daemon-hq594 Namespace:openshift-multus ID:592c0fe68eaae38ffff4788882a53055cedb618ddc573956838720a91f262ddf UID:50152296-9489-4fa8-aa42-7858debe1c08 NetNS:/var/run/netns/44c3e3ce-0d3a-4e35-9d33-048ce0d923c0 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.743512958Z" level=info msg="Adding pod openshift-multus_network-metrics-daemon-hq594 to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.765229166Z" level=error msg="Error stopping network on cleanup: failed to destroy network for pod sandbox k8s_network-check-target-xmq2g_openshift-network-diagnostics_f364a949-09d7-4cfb-83ff-e532b822a557_0(68c40ad6338d492b4aa5a2126d3afd81d13c48ead9d59f5b59bd6ce18b97e366): error removing pod openshift-network-diagnostics_network-check-target-xmq2g from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-network-diagnostics/network-check-target-xmq2g/f364a949-09d7-4cfb-83ff-e532b822a557]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" id=96524ba2-8daa-4a74-ab1c-222f3b9d9054 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.765278281Z" level=info msg="runSandbox: cleaning up namespaces after failing to run sandbox 68c40ad6338d492b4aa5a2126d3afd81d13c48ead9d59f5b59bd6ce18b97e366" id=96524ba2-8daa-4a74-ab1c-222f3b9d9054 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.776695115Z" level=info msg="Got pod network &{Name:prometheus-k8s-0 Namespace:openshift-monitoring ID:04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0 UID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 NetNS:/var/run/netns/358f9ab6-2e3e-47df-b3f3-eb6ac7120ea1 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.776730214Z" level=info msg="Adding pod openshift-monitoring_prometheus-k8s-0 to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.793045991Z" level=info msg="runSandbox: deleting pod ID a937bd9597c90954ad3041a6e3670bd2dfc453dbbd7b4d1b6d48e35dae7b83c0 from idIndex" id=0755a791-7501-4417-934f-53470d89bb87 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.793093718Z" level=info msg="runSandbox: removing pod sandbox a937bd9597c90954ad3041a6e3670bd2dfc453dbbd7b4d1b6d48e35dae7b83c0" id=0755a791-7501-4417-934f-53470d89bb87 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.793118645Z" level=info msg="runSandbox: deleting container ID from idIndex for sandbox a937bd9597c90954ad3041a6e3670bd2dfc453dbbd7b4d1b6d48e35dae7b83c0" id=0755a791-7501-4417-934f-53470d89bb87 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.793140849Z" level=info msg="runSandbox: unmounting shmPath for sandbox a937bd9597c90954ad3041a6e3670bd2dfc453dbbd7b4d1b6d48e35dae7b83c0" id=0755a791-7501-4417-934f-53470d89bb87 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.797123635Z" level=info msg="runSandbox: removing pod sandbox from storage: a937bd9597c90954ad3041a6e3670bd2dfc453dbbd7b4d1b6d48e35dae7b83c0" id=0755a791-7501-4417-934f-53470d89bb87 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.834964617Z" level=info msg="Got pod network &{Name:alertmanager-main-0 Namespace:openshift-monitoring ID:de5e12e9744f461b9b6c0146346e5117bb31a1f6538299e5dccde6d8250c4268 UID:05c9a6ea-140a-49a2-bde5-f220b05aa252 NetNS:/var/run/netns/6d71acfd-733e-4abe-9500-a64746e72ca6 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.835002077Z" level=info msg="Adding pod openshift-monitoring_alertmanager-main-0 to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.893781074Z" level=info msg="runSandbox: deleting pod ID 68c40ad6338d492b4aa5a2126d3afd81d13c48ead9d59f5b59bd6ce18b97e366 from idIndex" id=96524ba2-8daa-4a74-ab1c-222f3b9d9054 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.893843245Z" level=info msg="runSandbox: removing pod sandbox 68c40ad6338d492b4aa5a2126d3afd81d13c48ead9d59f5b59bd6ce18b97e366" id=96524ba2-8daa-4a74-ab1c-222f3b9d9054 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.893869426Z" level=info msg="runSandbox: deleting container ID from idIndex for sandbox 68c40ad6338d492b4aa5a2126d3afd81d13c48ead9d59f5b59bd6ce18b97e366" id=96524ba2-8daa-4a74-ab1c-222f3b9d9054 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.893922317Z" level=info msg="runSandbox: unmounting shmPath for sandbox 68c40ad6338d492b4aa5a2126d3afd81d13c48ead9d59f5b59bd6ce18b97e366" id=96524ba2-8daa-4a74-ab1c-222f3b9d9054 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.897524414Z" level=info msg="runSandbox: releasing container name: k8s_POD_kube-state-metrics-6798b548b9-xl88q_openshift-monitoring_2ecd2e2c-ab52-42e0-8fde-64e45804cc9e_0" id=0755a791-7501-4417-934f-53470d89bb87 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.897564194Z" level=info msg="runSandbox: releasing pod sandbox name: k8s_kube-state-metrics-6798b548b9-xl88q_openshift-monitoring_2ecd2e2c-ab52-42e0-8fde-64e45804cc9e_0" id=0755a791-7501-4417-934f-53470d89bb87 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.905819410Z" level=info msg="runSandbox: removing pod sandbox from storage: 68c40ad6338d492b4aa5a2126d3afd81d13c48ead9d59f5b59bd6ce18b97e366" id=96524ba2-8daa-4a74-ab1c-222f3b9d9054 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.909289 2927 remote_runtime.go:222] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-state-metrics-6798b548b9-xl88q_openshift-monitoring_2ecd2e2c-ab52-42e0-8fde-64e45804cc9e_0(a937bd9597c90954ad3041a6e3670bd2dfc453dbbd7b4d1b6d48e35dae7b83c0): error adding pod openshift-monitoring_kube-state-metrics-6798b548b9-xl88q to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.909360 2927 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-state-metrics-6798b548b9-xl88q_openshift-monitoring_2ecd2e2c-ab52-42e0-8fde-64e45804cc9e_0(a937bd9597c90954ad3041a6e3670bd2dfc453dbbd7b4d1b6d48e35dae7b83c0): error adding pod openshift-monitoring_kube-state-metrics-6798b548b9-xl88q to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.909397 2927 kuberuntime_manager.go:772] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-state-metrics-6798b548b9-xl88q_openshift-monitoring_2ecd2e2c-ab52-42e0-8fde-64e45804cc9e_0(a937bd9597c90954ad3041a6e3670bd2dfc453dbbd7b4d1b6d48e35dae7b83c0): error adding pod openshift-monitoring_kube-state-metrics-6798b548b9-xl88q to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:35.909470 2927 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"kube-state-metrics-6798b548b9-xl88q_openshift-monitoring(2ecd2e2c-ab52-42e0-8fde-64e45804cc9e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"kube-state-metrics-6798b548b9-xl88q_openshift-monitoring(2ecd2e2c-ab52-42e0-8fde-64e45804cc9e)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-state-metrics-6798b548b9-xl88q_openshift-monitoring_2ecd2e2c-ab52-42e0-8fde-64e45804cc9e_0(a937bd9597c90954ad3041a6e3670bd2dfc453dbbd7b4d1b6d48e35dae7b83c0): error adding pod openshift-monitoring_kube-state-metrics-6798b548b9-xl88q to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus\\\" name=\\\"multus-cni-network\\\" failed (add): Multus: [openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition\"" pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" podUID=2ecd2e2c-ab52-42e0-8fde-64e45804cc9e Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.950925736Z" level=info msg="Got pod network &{Name:prometheus-adapter-5894cc86c7-qhmbh Namespace:openshift-monitoring ID:fe3e7b123db46a0b88eddb5d106d43267fc1d0c1707470c0568f2b211ee6ba55 UID:a40bf20d-67a4-45e7-b701-ca5a75376c1f NetNS:/var/run/netns/d1d94e66-dbb6-4156-bf1f-528844201625 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:35.950964551Z" level=info msg="Adding pod openshift-monitoring_prometheus-adapter-5894cc86c7-qhmbh to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.044273084Z" level=info msg="runSandbox: releasing container name: k8s_POD_network-check-target-xmq2g_openshift-network-diagnostics_f364a949-09d7-4cfb-83ff-e532b822a557_0" id=96524ba2-8daa-4a74-ab1c-222f3b9d9054 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.044316970Z" level=info msg="runSandbox: releasing pod sandbox name: k8s_network-check-target-xmq2g_openshift-network-diagnostics_f364a949-09d7-4cfb-83ff-e532b822a557_0" id=96524ba2-8daa-4a74-ab1c-222f3b9d9054 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:36.047900 2927 remote_runtime.go:222] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_network-check-target-xmq2g_openshift-network-diagnostics_f364a949-09d7-4cfb-83ff-e532b822a557_0(68c40ad6338d492b4aa5a2126d3afd81d13c48ead9d59f5b59bd6ce18b97e366): error adding pod openshift-network-diagnostics_network-check-target-xmq2g to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-network-diagnostics/network-check-target-xmq2g/f364a949-09d7-4cfb-83ff-e532b822a557]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:36.047993 2927 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_network-check-target-xmq2g_openshift-network-diagnostics_f364a949-09d7-4cfb-83ff-e532b822a557_0(68c40ad6338d492b4aa5a2126d3afd81d13c48ead9d59f5b59bd6ce18b97e366): error adding pod openshift-network-diagnostics_network-check-target-xmq2g to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-network-diagnostics/network-check-target-xmq2g/f364a949-09d7-4cfb-83ff-e532b822a557]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:36.048083 2927 kuberuntime_manager.go:772] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_network-check-target-xmq2g_openshift-network-diagnostics_f364a949-09d7-4cfb-83ff-e532b822a557_0(68c40ad6338d492b4aa5a2126d3afd81d13c48ead9d59f5b59bd6ce18b97e366): error adding pod openshift-network-diagnostics_network-check-target-xmq2g to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-network-diagnostics/network-check-target-xmq2g/f364a949-09d7-4cfb-83ff-e532b822a557]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:36.048177 2927 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"network-check-target-xmq2g_openshift-network-diagnostics(f364a949-09d7-4cfb-83ff-e532b822a557)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"network-check-target-xmq2g_openshift-network-diagnostics(f364a949-09d7-4cfb-83ff-e532b822a557)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_network-check-target-xmq2g_openshift-network-diagnostics_f364a949-09d7-4cfb-83ff-e532b822a557_0(68c40ad6338d492b4aa5a2126d3afd81d13c48ead9d59f5b59bd6ce18b97e366): error adding pod openshift-network-diagnostics_network-check-target-xmq2g to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus\\\" name=\\\"multus-cni-network\\\" failed (add): Multus: [openshift-network-diagnostics/network-check-target-xmq2g/f364a949-09d7-4cfb-83ff-e532b822a557]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition\"" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.051416266Z" level=info msg="Got pod network &{Name:prometheus-operator-admission-webhook-598468bb49-94dsn Namespace:openshift-monitoring ID:90d2b912f09210aef9996180fedf7e65296d392716fb68985112976a13c8d29b UID:88745ab7-efe7-42c0-a2e8-75591518333d NetNS:/var/run/netns/58506b8e-16a1-4c50-8f8b-a8b3adf8da54 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.051455560Z" level=info msg="Adding pod openshift-monitoring_prometheus-operator-admission-webhook-598468bb49-94dsn to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.051422749Z" level=info msg="Got pod network &{Name:router-default-5cf6bbdff9-tnffs Namespace:openshift-ingress ID:68d9bc9c6a9bcf27e7bcf19bef0f4b3c6707eefbb0ba0e388f51a8e663a95ad7 UID:c3b10ff4-a243-4a4e-b27d-f613e258666d NetNS:/var/run/netns/b8f70c62-7b79-4f76-b3e5-1851b2785ffa Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.051689002Z" level=info msg="Adding pod openshift-ingress_router-default-5cf6bbdff9-tnffs to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.080216732Z" level=info msg="Got pod network &{Name:ingress-canary-8xrbh Namespace:openshift-ingress-canary ID:f226c6f82e17c96c33e4b6a3a2d81c74b2301377f9323d0a64195f4af1cc5679 UID:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748 NetNS:/var/run/netns/cfdfca99-9582-448d-b3cd-a75ebd263536 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.080254552Z" level=info msg="Adding pod openshift-ingress-canary_ingress-canary-8xrbh to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.406329596Z" level=error msg="Error stopping network on cleanup: failed to destroy network for pod sandbox k8s_openshift-state-metrics-6f5dcdf79b-wq256_openshift-monitoring_9de5de36-9b80-4a0b-a615-f0f9ccda2ff6_0(afb580e701142b75f062c265b00883e2d5567b37000d823ade7ce8eb8a057b02): error removing pod openshift-monitoring_openshift-state-metrics-6f5dcdf79b-wq256 from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" id=ff04d155-666b-4a8f-ba4b-1d1da011fa66 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.406385961Z" level=info msg="runSandbox: cleaning up namespaces after failing to run sandbox afb580e701142b75f062c265b00883e2d5567b37000d823ade7ce8eb8a057b02" id=ff04d155-666b-4a8f-ba4b-1d1da011fa66 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:36.581923 2927 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/telemeter-client-5598c645c4-mftdv] Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: I0105 09:29:36.148018 6250 ovs.go:90] Maximum command line arguments set to: 191102 Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: 2023-01-05T09:29:36Z [verbose] Add: openshift-monitoring:telemeter-client-5598c645c4-mftdv:fc45e847-8f55-4c48-8c48-4bfee2419dca:ovn-kubernetes(ovn-kubernetes):eth0 {"cniVersion":"0.4.0","interfaces":[{"name":"f1e02d518b5e62e","mac":"02:d2:d9:66:b4:c8"},{"name":"eth0","mac":"0a:58:0a:83:00:0b","sandbox":"/var/run/netns/04b13e36-0acc-4f75-9752-0ebb73acd4ae"}],"ips":[{"version":"4","interface":1,"address":"10.131.0.11/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: I0105 09:29:36.547709 6239 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"telemeter-client-5598c645c4-mftdv", UID:"fc45e847-8f55-4c48-8c48-4bfee2419dca", APIVersion:"v1", ResourceVersion:"90070", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.11/23] from ovn-kubernetes Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.586546206Z" level=info msg="Got pod network &{Name:telemeter-client-5598c645c4-mftdv Namespace:openshift-monitoring ID:f1e02d518b5e62edc187ed7cc730fab2279f2eb5384e3f3ffd25d3a014b31d4b UID:fc45e847-8f55-4c48-8c48-4bfee2419dca NetNS:/var/run/netns/04b13e36-0acc-4f75-9752-0ebb73acd4ae Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.586724618Z" level=info msg="Checking pod openshift-monitoring_telemeter-client-5598c645c4-mftdv for CNI network multus-cni-network (type=multus)" Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.636433197Z" level=info msg="Ran pod sandbox f1e02d518b5e62edc187ed7cc730fab2279f2eb5384e3f3ffd25d3a014b31d4b with infra container: openshift-monitoring/telemeter-client-5598c645c4-mftdv/POD" id=2240f64b-1380-422d-8c5e-4835cf6a133c name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.653808590Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:85f671344a3bc1584db4caed39939b6dd55644a5bd806e118527f31dac91252e" id=507d24c8-11eb-48f1-b701-db9ab736311a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.654074787Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b66b2c4519cdd77137d5eaa713199fe202373c0d1e029b31472b9fc8a68a2aa7,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:85f671344a3bc1584db4caed39939b6dd55644a5bd806e118527f31dac91252e],Size_:338794817,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=507d24c8-11eb-48f1-b701-db9ab736311a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.674234489Z" level=info msg="runSandbox: deleting pod ID afb580e701142b75f062c265b00883e2d5567b37000d823ade7ce8eb8a057b02 from idIndex" id=ff04d155-666b-4a8f-ba4b-1d1da011fa66 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.674350802Z" level=info msg="runSandbox: removing pod sandbox afb580e701142b75f062c265b00883e2d5567b37000d823ade7ce8eb8a057b02" id=ff04d155-666b-4a8f-ba4b-1d1da011fa66 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.674382344Z" level=info msg="runSandbox: deleting container ID from idIndex for sandbox afb580e701142b75f062c265b00883e2d5567b37000d823ade7ce8eb8a057b02" id=ff04d155-666b-4a8f-ba4b-1d1da011fa66 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.674410094Z" level=info msg="runSandbox: unmounting shmPath for sandbox afb580e701142b75f062c265b00883e2d5567b37000d823ade7ce8eb8a057b02" id=ff04d155-666b-4a8f-ba4b-1d1da011fa66 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.674679970Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:85f671344a3bc1584db4caed39939b6dd55644a5bd806e118527f31dac91252e" id=e65eac9a-8535-425a-a005-8effd26390a2 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.675001753Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b66b2c4519cdd77137d5eaa713199fe202373c0d1e029b31472b9fc8a68a2aa7,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:85f671344a3bc1584db4caed39939b6dd55644a5bd806e118527f31dac91252e],Size_:338794817,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=e65eac9a-8535-425a-a005-8effd26390a2 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.681279632Z" level=info msg="Creating container: openshift-monitoring/telemeter-client-5598c645c4-mftdv/telemeter-client" id=c4c55c63-e0f8-47b4-81b9-f3f0c43b033d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.681412203Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.690749067Z" level=info msg="runSandbox: removing pod sandbox from storage: afb580e701142b75f062c265b00883e2d5567b37000d823ade7ce8eb8a057b02" id=ff04d155-666b-4a8f-ba4b-1d1da011fa66 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:36.712436 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:36.712652 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" event=&{ID:fc45e847-8f55-4c48-8c48-4bfee2419dca Type:ContainerStarted Data:f1e02d518b5e62edc187ed7cc730fab2279f2eb5384e3f3ffd25d3a014b31d4b} Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:36.713087 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-2pfzf" Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:36.713582 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.714207209Z" level=info msg="Running pod sandbox: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/POD" id=3b934532-6ee6-4d74-a5b1-22db4939f2d8 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.714262466Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.714399121Z" level=info msg="Running pod sandbox: openshift-network-diagnostics/network-check-target-xmq2g/POD" id=b6c480b2-0246-48e3-b96d-279deb94c18e name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.714429064Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.714505990Z" level=info msg="Running pod sandbox: openshift-dns/dns-default-2pfzf/POD" id=349ce886-a7b4-4d61-8900-38fba3bfc31d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.714530276Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:36.893008 2927 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-multus/network-metrics-daemon-hq594] Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.907985922Z" level=info msg="runSandbox: releasing container name: k8s_POD_openshift-state-metrics-6f5dcdf79b-wq256_openshift-monitoring_9de5de36-9b80-4a0b-a615-f0f9ccda2ff6_0" id=ff04d155-666b-4a8f-ba4b-1d1da011fa66 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.908032611Z" level=info msg="runSandbox: releasing pod sandbox name: k8s_openshift-state-metrics-6f5dcdf79b-wq256_openshift-monitoring_9de5de36-9b80-4a0b-a615-f0f9ccda2ff6_0" id=ff04d155-666b-4a8f-ba4b-1d1da011fa66 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: I0105 09:29:36.471707 6260 ovs.go:90] Maximum command line arguments set to: 191102 Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: 2023-01-05T09:29:36Z [verbose] Add: openshift-multus:network-metrics-daemon-hq594:50152296-9489-4fa8-aa42-7858debe1c08:ovn-kubernetes(ovn-kubernetes):eth0 {"cniVersion":"0.4.0","interfaces":[{"name":"592c0fe68eaae38","mac":"16:90:8b:a5:a9:8f"},{"name":"eth0","mac":"0a:58:0a:83:00:0c","sandbox":"/var/run/netns/44c3e3ce-0d3a-4e35-9d33-048ce0d923c0"}],"ips":[{"version":"4","interface":1,"address":"10.131.0.12/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: I0105 09:29:36.854295 6243 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-multus", Name:"network-metrics-daemon-hq594", UID:"50152296-9489-4fa8-aa42-7858debe1c08", APIVersion:"v1", ResourceVersion:"90087", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.12/23] from ovn-kubernetes Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:36.919990 2927 remote_runtime.go:222] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-state-metrics-6f5dcdf79b-wq256_openshift-monitoring_9de5de36-9b80-4a0b-a615-f0f9ccda2ff6_0(afb580e701142b75f062c265b00883e2d5567b37000d823ade7ce8eb8a057b02): error adding pod openshift-monitoring_openshift-state-metrics-6f5dcdf79b-wq256 to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:36.920058 2927 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-state-metrics-6f5dcdf79b-wq256_openshift-monitoring_9de5de36-9b80-4a0b-a615-f0f9ccda2ff6_0(afb580e701142b75f062c265b00883e2d5567b37000d823ade7ce8eb8a057b02): error adding pod openshift-monitoring_openshift-state-metrics-6f5dcdf79b-wq256 to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:36.920104 2927 kuberuntime_manager.go:772] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-state-metrics-6f5dcdf79b-wq256_openshift-monitoring_9de5de36-9b80-4a0b-a615-f0f9ccda2ff6_0(afb580e701142b75f062c265b00883e2d5567b37000d823ade7ce8eb8a057b02): error adding pod openshift-monitoring_openshift-state-metrics-6f5dcdf79b-wq256 to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:36.920190 2927 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"openshift-state-metrics-6f5dcdf79b-wq256_openshift-monitoring(9de5de36-9b80-4a0b-a615-f0f9ccda2ff6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"openshift-state-metrics-6f5dcdf79b-wq256_openshift-monitoring(9de5de36-9b80-4a0b-a615-f0f9ccda2ff6)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-state-metrics-6f5dcdf79b-wq256_openshift-monitoring_9de5de36-9b80-4a0b-a615-f0f9ccda2ff6_0(afb580e701142b75f062c265b00883e2d5567b37000d823ade7ce8eb8a057b02): error adding pod openshift-monitoring_openshift-state-metrics-6f5dcdf79b-wq256 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus\\\" name=\\\"multus-cni-network\\\" failed (add): Multus: [openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition\"" pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" podUID=9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.921752167Z" level=info msg="Got pod network &{Name:network-metrics-daemon-hq594 Namespace:openshift-multus ID:592c0fe68eaae38ffff4788882a53055cedb618ddc573956838720a91f262ddf UID:50152296-9489-4fa8-aa42-7858debe1c08 NetNS:/var/run/netns/44c3e3ce-0d3a-4e35-9d33-048ce0d923c0 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.921958987Z" level=info msg="Checking pod openshift-multus_network-metrics-daemon-hq594 for CNI network multus-cni-network (type=multus)" Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.930560201Z" level=info msg="Got pod network &{Name:network-check-target-xmq2g Namespace:openshift-network-diagnostics ID:8a313b3accfad3269666173b7b1c660df2741df597e595c3c47a5ca7eb07e228 UID:f364a949-09d7-4cfb-83ff-e532b822a557 NetNS:/var/run/netns/c226eea0-515f-4498-858c-3c6f29b4b2db Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:36.930595258Z" level=info msg="Adding pod openshift-network-diagnostics_network-check-target-xmq2g to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: I0105 09:29:36.649390 6268 ovs.go:90] Maximum command line arguments set to: 191102 Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: 2023-01-05T09:29:36Z [verbose] Add: openshift-monitoring:prometheus-k8s-0:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879:ovn-kubernetes(ovn-kubernetes):eth0 {"cniVersion":"0.4.0","interfaces":[{"name":"04ef5a2380724b3","mac":"42:61:7c:97:ac:c0"},{"name":"eth0","mac":"0a:58:0a:83:00:06","sandbox":"/var/run/netns/358f9ab6-2e3e-47df-b3f3-eb6ac7120ea1"}],"ips":[{"version":"4","interface":1,"address":"10.131.0.6/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: I0105 09:29:36.926846 6255 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"prometheus-k8s-0", UID:"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879", APIVersion:"v1", ResourceVersion:"89990", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.6/23] from ovn-kubernetes Jan 05 09:29:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:29:36.975288 2927 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50152296_9489_4fa8_aa42_7858debe1c08.slice/crio-592c0fe68eaae38ffff4788882a53055cedb618ddc573956838720a91f262ddf.scope WatchSource:0}: Error finding container 592c0fe68eaae38ffff4788882a53055cedb618ddc573956838720a91f262ddf: Status 404 returned error can't find the container with id 592c0fe68eaae38ffff4788882a53055cedb618ddc573956838720a91f262ddf Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.037722022Z" level=info msg="Got pod network &{Name:prometheus-k8s-0 Namespace:openshift-monitoring ID:04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0 UID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 NetNS:/var/run/netns/358f9ab6-2e3e-47df-b3f3-eb6ac7120ea1 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.038022736Z" level=info msg="Checking pod openshift-monitoring_prometheus-k8s-0 for CNI network multus-cni-network (type=multus)" Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.051142703Z" level=info msg="Ran pod sandbox 592c0fe68eaae38ffff4788882a53055cedb618ddc573956838720a91f262ddf with infra container: openshift-multus/network-metrics-daemon-hq594/POD" id=3735d57a-2221-4f15-af34-ce6f61dcab84 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:37.068744 2927 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/prometheus-k8s-0] Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.069500756Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9e33552d1aaea9cf6a301fa438cfc8305826c5ddb58ac8437d884e55e71c6e44" id=ee72fb07-38ea-4006-acff-c0de7f82f96c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.069718740Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:7b87fc786118817b8913f28a61e01db37b08380f754bbaacf425508b208632a0,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9e33552d1aaea9cf6a301fa438cfc8305826c5ddb58ac8437d884e55e71c6e44],Size_:312894682,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=ee72fb07-38ea-4006-acff-c0de7f82f96c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.078028620Z" level=info msg="Got pod network &{Name:kube-state-metrics-6798b548b9-xl88q Namespace:openshift-monitoring ID:9c318dae338ceca029f5b8d71eed7d2265d6188623995ec40e8d80da65a8d993 UID:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e NetNS:/var/run/netns/8598084b-1708-4476-81ee-e77d49e075ea Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.078058628Z" level=info msg="Adding pod openshift-monitoring_kube-state-metrics-6798b548b9-xl88q to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.080145455Z" level=info msg="Ran pod sandbox 04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0 with infra container: openshift-monitoring/prometheus-k8s-0/POD" id=e2aefa06-9295-4ff6-9b6b-ff22dd9746c2 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.083315094Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9e33552d1aaea9cf6a301fa438cfc8305826c5ddb58ac8437d884e55e71c6e44" id=3ec4de31-fd7b-41a7-8e44-fef437560183 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.088296962Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=75e7a239-36e3-46b8-9fcc-6865118ec411 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.098592759Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=75e7a239-36e3-46b8-9fcc-6865118ec411 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.098784556Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:7b87fc786118817b8913f28a61e01db37b08380f754bbaacf425508b208632a0,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9e33552d1aaea9cf6a301fa438cfc8305826c5ddb58ac8437d884e55e71c6e44],Size_:312894682,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=3ec4de31-fd7b-41a7-8e44-fef437560183 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.107074893Z" level=info msg="Creating container: openshift-multus/network-metrics-daemon-hq594/network-metrics-daemon" id=da66efb7-2d46-4f8c-ad06-bfe86e8d328c name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.107186678Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.109478659Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=e80751a2-c6cf-4922-bf2b-5972bdfa3e9b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.261982260Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=e80751a2-c6cf-4922-bf2b-5972bdfa3e9b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.287470919Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-0/init-config-reloader" id=0ab117f2-dbca-4820-a442-d3bfac49c381 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.287599703Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.456527034Z" level=info msg="Got pod network &{Name:dns-default-2pfzf Namespace:openshift-dns ID:238eef320be8a6fa61ec83ea4bf9cdc139cde1247f51dbe0bd0040a8e9c70088 UID:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 NetNS:/var/run/netns/8e512d8a-b52a-45da-a484-9148509a8cea Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.456570845Z" level=info msg="Adding pod openshift-dns_dns-default-2pfzf to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:37.646750 2927 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh] Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: I0105 09:29:37.105813 6298 ovs.go:90] Maximum command line arguments set to: 191102 Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: 2023-01-05T09:29:37Z [verbose] Add: openshift-monitoring:prometheus-adapter-5894cc86c7-qhmbh:a40bf20d-67a4-45e7-b701-ca5a75376c1f:ovn-kubernetes(ovn-kubernetes):eth0 {"cniVersion":"0.4.0","interfaces":[{"name":"fe3e7b123db46a0","mac":"2a:1e:89:7e:f4:65"},{"name":"eth0","mac":"0a:58:0a:83:00:03","sandbox":"/var/run/netns/d1d94e66-dbb6-4156-bf1f-528844201625"}],"ips":[{"version":"4","interface":1,"address":"10.131.0.3/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: I0105 09:29:37.599430 6282 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"prometheus-adapter-5894cc86c7-qhmbh", UID:"a40bf20d-67a4-45e7-b701-ca5a75376c1f", APIVersion:"v1", ResourceVersion:"89959", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.3/23] from ovn-kubernetes Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.679624765Z" level=info msg="Got pod network &{Name:prometheus-adapter-5894cc86c7-qhmbh Namespace:openshift-monitoring ID:fe3e7b123db46a0b88eddb5d106d43267fc1d0c1707470c0568f2b211ee6ba55 UID:a40bf20d-67a4-45e7-b701-ca5a75376c1f NetNS:/var/run/netns/d1d94e66-dbb6-4156-bf1f-528844201625 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.679819443Z" level=info msg="Checking pod openshift-monitoring_prometheus-adapter-5894cc86c7-qhmbh for CNI network multus-cni-network (type=multus)" Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:29:37.736929 2927 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda40bf20d_67a4_45e7_b701_ca5a75376c1f.slice/crio-fe3e7b123db46a0b88eddb5d106d43267fc1d0c1707470c0568f2b211ee6ba55.scope WatchSource:0}: Error finding container fe3e7b123db46a0b88eddb5d106d43267fc1d0c1707470c0568f2b211ee6ba55: Status 404 returned error can't find the container with id fe3e7b123db46a0b88eddb5d106d43267fc1d0c1707470c0568f2b211ee6ba55 Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.737227838Z" level=info msg="Ran pod sandbox fe3e7b123db46a0b88eddb5d106d43267fc1d0c1707470c0568f2b211ee6ba55 with infra container: openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh/POD" id=f8a4f035-df5c-4a67-95f7-6afbfb299b6e name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.769065171Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9696f8f231d49aadf63ab980ec5c17c4fd87011930daecd30cc9e4ee102f1aeb" id=d85b038a-ecde-45a3-8a9b-c769a13dfb64 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.769356876Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:ea0b364e3a8a4e4b3f4ab3288e42c5be8021187d32791029f9cb9a500c8da100,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9696f8f231d49aadf63ab980ec5c17c4fd87011930daecd30cc9e4ee102f1aeb],Size_:402351209,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=d85b038a-ecde-45a3-8a9b-c769a13dfb64 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.802503677Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9696f8f231d49aadf63ab980ec5c17c4fd87011930daecd30cc9e4ee102f1aeb" id=5045ce40-1745-4b52-91ad-ec3f13d296c8 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.802780268Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:ea0b364e3a8a4e4b3f4ab3288e42c5be8021187d32791029f9cb9a500c8da100,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9696f8f231d49aadf63ab980ec5c17c4fd87011930daecd30cc9e4ee102f1aeb],Size_:402351209,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=5045ce40-1745-4b52-91ad-ec3f13d296c8 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.809205978Z" level=info msg="Creating container: openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh/prometheus-adapter" id=16a2b9b7-bb41-4c14-8921-d3b711789de4 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.809328289Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:37.814394 2927 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/alertmanager-main-0] Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:37.825787 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event=&{ID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Type:ContainerStarted Data:04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0} Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:37.866448 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:37.866667 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hq594" event=&{ID:50152296-9489-4fa8-aa42-7858debe1c08 Type:ContainerStarted Data:592c0fe68eaae38ffff4788882a53055cedb618ddc573956838720a91f262ddf} Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.868423178Z" level=info msg="Running pod sandbox: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/POD" id=fc9103cb-ce68-45a9-b08a-00886595b7e5 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.868492797Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: I0105 09:29:37.190691 6281 ovs.go:90] Maximum command line arguments set to: 191102 Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: 2023-01-05T09:29:37Z [verbose] Add: openshift-monitoring:alertmanager-main-0:05c9a6ea-140a-49a2-bde5-f220b05aa252:ovn-kubernetes(ovn-kubernetes):eth0 {"cniVersion":"0.4.0","interfaces":[{"name":"de5e12e9744f461","mac":"da:56:09:86:05:e2"},{"name":"eth0","mac":"0a:58:0a:83:00:09","sandbox":"/var/run/netns/6d71acfd-733e-4abe-9500-a64746e72ca6"}],"ips":[{"version":"4","interface":1,"address":"10.131.0.9/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: I0105 09:29:37.734242 6266 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"alertmanager-main-0", UID:"05c9a6ea-140a-49a2-bde5-f220b05aa252", APIVersion:"v1", ResourceVersion:"90065", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.9/23] from ovn-kubernetes Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.872702218Z" level=info msg="Got pod network &{Name:alertmanager-main-0 Namespace:openshift-monitoring ID:de5e12e9744f461b9b6c0146346e5117bb31a1f6538299e5dccde6d8250c4268 UID:05c9a6ea-140a-49a2-bde5-f220b05aa252 NetNS:/var/run/netns/6d71acfd-733e-4abe-9500-a64746e72ca6 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.872850450Z" level=info msg="Checking pod openshift-monitoring_alertmanager-main-0 for CNI network multus-cni-network (type=multus)" Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.930570558Z" level=info msg="Ran pod sandbox de5e12e9744f461b9b6c0146346e5117bb31a1f6538299e5dccde6d8250c4268 with infra container: openshift-monitoring/alertmanager-main-0/POD" id=40151f20-e5d6-484a-a00f-ed789a9bb157 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.972556814Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e6173b5d77def9e8058c7ff82f1ad65f537d1a81f88fcdabe6d8e30bff49ca89" id=5e7d06d7-f9a2-4761-a8fa-f3c2d264ba4e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:37.972792215Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:d168c738461c25cc0c3fa96cfcdf2dc337ca08f5bf3a6b550fcb29a59ea39270,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e6173b5d77def9e8058c7ff82f1ad65f537d1a81f88fcdabe6d8e30bff49ca89],Size_:333780768,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=5e7d06d7-f9a2-4761-a8fa-f3c2d264ba4e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.002279751Z" level=info msg="Created container 474834dd0b54abe011163e7458744c9ca3a52b894e6a07cfb86b009916a74367: openshift-monitoring/telemeter-client-5598c645c4-mftdv/telemeter-client" id=c4c55c63-e0f8-47b4-81b9-f3f0c43b033d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.013134126Z" level=info msg="Starting container: 474834dd0b54abe011163e7458744c9ca3a52b894e6a07cfb86b009916a74367" id=48af5a0f-0cd5-4682-8dc3-4d89b1730aa0 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.061776290Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e6173b5d77def9e8058c7ff82f1ad65f537d1a81f88fcdabe6d8e30bff49ca89" id=9fd370ad-cbec-4cf3-84da-3de4b14b63f9 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.062072332Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:d168c738461c25cc0c3fa96cfcdf2dc337ca08f5bf3a6b550fcb29a59ea39270,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e6173b5d77def9e8058c7ff82f1ad65f537d1a81f88fcdabe6d8e30bff49ca89],Size_:333780768,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=9fd370ad-cbec-4cf3-84da-3de4b14b63f9 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.072297776Z" level=info msg="Creating container: openshift-monitoring/alertmanager-main-0/alertmanager" id=f18bf7cd-9b19-407e-b0b2-c71686dccc94 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.072426120Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.191772375Z" level=info msg="Got pod network &{Name:openshift-state-metrics-6f5dcdf79b-wq256 Namespace:openshift-monitoring ID:318b2d4d118085d70f305dfbfb7875ec65ba1a3f9539d6e1f29f152cb2d8af82 UID:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 NetNS:/var/run/netns/d458dea1-0719-4a13-bc09-29a80e493860 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.191813141Z" level=info msg="Adding pod openshift-monitoring_openshift-state-metrics-6f5dcdf79b-wq256 to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:38.225491 2927 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-ingress-canary/ingress-canary-8xrbh] Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: I0105 09:29:37.578201 6350 ovs.go:90] Maximum command line arguments set to: 191102 Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: 2023-01-05T09:29:38Z [verbose] Add: openshift-ingress-canary:ingress-canary-8xrbh:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748:ovn-kubernetes(ovn-kubernetes):eth0 {"cniVersion":"0.4.0","interfaces":[{"name":"f226c6f82e17c96","mac":"0e:df:b9:24:3a:9e"},{"name":"eth0","mac":"0a:58:0a:83:00:0e","sandbox":"/var/run/netns/cfdfca99-9582-448d-b3cd-a75ebd263536"}],"ips":[{"version":"4","interface":1,"address":"10.131.0.14/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: I0105 09:29:38.184710 6311 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-ingress-canary", Name:"ingress-canary-8xrbh", UID:"e04a47b4-aa8e-4d4e-9d45-aa8037dcb748", APIVersion:"v1", ResourceVersion:"90092", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.14/23] from ovn-kubernetes Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.243127540Z" level=info msg="Got pod network &{Name:ingress-canary-8xrbh Namespace:openshift-ingress-canary ID:f226c6f82e17c96c33e4b6a3a2d81c74b2301377f9323d0a64195f4af1cc5679 UID:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748 NetNS:/var/run/netns/cfdfca99-9582-448d-b3cd-a75ebd263536 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.243301226Z" level=info msg="Checking pod openshift-ingress-canary_ingress-canary-8xrbh for CNI network multus-cni-network (type=multus)" Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:29:38.262169 2927 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode04a47b4_aa8e_4d4e_9d45_aa8037dcb748.slice/crio-f226c6f82e17c96c33e4b6a3a2d81c74b2301377f9323d0a64195f4af1cc5679.scope WatchSource:0}: Error finding container f226c6f82e17c96c33e4b6a3a2d81c74b2301377f9323d0a64195f4af1cc5679: Status 404 returned error can't find the container with id f226c6f82e17c96c33e4b6a3a2d81c74b2301377f9323d0a64195f4af1cc5679 Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.270057641Z" level=info msg="Ran pod sandbox f226c6f82e17c96c33e4b6a3a2d81c74b2301377f9323d0a64195f4af1cc5679 with infra container: openshift-ingress-canary/ingress-canary-8xrbh/POD" id=900808c6-6154-478e-8e45-2a15867c06a1 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:38.278353 2927 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn] Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: I0105 09:29:37.585229 6327 ovs.go:90] Maximum command line arguments set to: 191102 Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: 2023-01-05T09:29:38Z [verbose] Add: openshift-monitoring:prometheus-operator-admission-webhook-598468bb49-94dsn:88745ab7-efe7-42c0-a2e8-75591518333d:ovn-kubernetes(ovn-kubernetes):eth0 {"cniVersion":"0.4.0","interfaces":[{"name":"90d2b912f09210a","mac":"42:6b:c3:63:5c:e0"},{"name":"eth0","mac":"0a:58:0a:83:00:0f","sandbox":"/var/run/netns/58506b8e-16a1-4c50-8f8b-a8b3adf8da54"}],"ips":[{"version":"4","interface":1,"address":"10.131.0.15/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: I0105 09:29:38.198442 6300 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"prometheus-operator-admission-webhook-598468bb49-94dsn", UID:"88745ab7-efe7-42c0-a2e8-75591518333d", APIVersion:"v1", ResourceVersion:"90110", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.15/23] from ovn-kubernetes Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.368307054Z" level=info msg="Got pod network &{Name:prometheus-operator-admission-webhook-598468bb49-94dsn Namespace:openshift-monitoring ID:90d2b912f09210aef9996180fedf7e65296d392716fb68985112976a13c8d29b UID:88745ab7-efe7-42c0-a2e8-75591518333d NetNS:/var/run/netns/58506b8e-16a1-4c50-8f8b-a8b3adf8da54 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.368465471Z" level=info msg="Checking pod openshift-monitoring_prometheus-operator-admission-webhook-598468bb49-94dsn for CNI network multus-cni-network (type=multus)" Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.371162672Z" level=info msg="Started container" PID=6431 containerID=474834dd0b54abe011163e7458744c9ca3a52b894e6a07cfb86b009916a74367 description=openshift-monitoring/telemeter-client-5598c645c4-mftdv/telemeter-client id=48af5a0f-0cd5-4682-8dc3-4d89b1730aa0 name=/runtime.v1.RuntimeService/StartContainer sandboxID=f1e02d518b5e62edc187ed7cc730fab2279f2eb5384e3f3ffd25d3a014b31d4b Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.382609823Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:5974989cda2655deb3052d601e7406eaf3caee0e7999a6eff251760aac19c199" id=52978e33-a153-473d-9916-59a0ab24a357 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.382833425Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:c34c97b229e46faa94bab6ad2559fd683bd036df06031c9097dd27f7daaf96c7,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:5974989cda2655deb3052d601e7406eaf3caee0e7999a6eff251760aac19c199],Size_:417780553,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=52978e33-a153-473d-9916-59a0ab24a357 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.388307011Z" level=info msg="Created container dbf94c2e9e290b7a0ff09b82e88404bfcefb5c50d5c57120c11ad6bd75e3d27b: openshift-multus/network-metrics-daemon-hq594/network-metrics-daemon" id=da66efb7-2d46-4f8c-ad06-bfe86e8d328c name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.389739240Z" level=info msg="Ran pod sandbox 90d2b912f09210aef9996180fedf7e65296d392716fb68985112976a13c8d29b with infra container: openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn/POD" id=f9385c94-5f07-474c-a4be-b65b96cb4672 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:29:38.421589 2927 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88745ab7_efe7_42c0_a2e8_75591518333d.slice/crio-90d2b912f09210aef9996180fedf7e65296d392716fb68985112976a13c8d29b.scope WatchSource:0}: Error finding container 90d2b912f09210aef9996180fedf7e65296d392716fb68985112976a13c8d29b: Status 404 returned error can't find the container with id 90d2b912f09210aef9996180fedf7e65296d392716fb68985112976a13c8d29b Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.426977114Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:5974989cda2655deb3052d601e7406eaf3caee0e7999a6eff251760aac19c199" id=19c81a0a-d2d5-4190-9ecc-dc406f1c8774 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.427220750Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:c34c97b229e46faa94bab6ad2559fd683bd036df06031c9097dd27f7daaf96c7,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:5974989cda2655deb3052d601e7406eaf3caee0e7999a6eff251760aac19c199],Size_:417780553,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=19c81a0a-d2d5-4190-9ecc-dc406f1c8774 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.427364231Z" level=info msg="Starting container: dbf94c2e9e290b7a0ff09b82e88404bfcefb5c50d5c57120c11ad6bd75e3d27b" id=7be066f8-5e0b-4f5b-997a-f3e6ab8d460f name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.428781917Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:dbf46164dcc3b2b6ef5df5d043ddc4fd859fd2c73c4862808424b3df6a1e27e7" id=aec7e27e-428d-412c-96ec-49dfd9e5444a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.429003377Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:20e9403b34a0d9d25151a4ddbee3f69b99894fd8e058e6cb7310fe95080ecb03,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:dbf46164dcc3b2b6ef5df5d043ddc4fd859fd2c73c4862808424b3df6a1e27e7],Size_:374119253,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=aec7e27e-428d-412c-96ec-49dfd9e5444a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.429180702Z" level=info msg="Creating container: openshift-ingress-canary/ingress-canary-8xrbh/serve-healthcheck-canary" id=7a2a51dc-cd2b-4c0f-8027-8a084dcb96bd name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.429295683Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.432844618Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:dbf46164dcc3b2b6ef5df5d043ddc4fd859fd2c73c4862808424b3df6a1e27e7" id=844df08c-77a4-4ae3-83f6-20af3d43c259 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.448680362Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:20e9403b34a0d9d25151a4ddbee3f69b99894fd8e058e6cb7310fe95080ecb03,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:dbf46164dcc3b2b6ef5df5d043ddc4fd859fd2c73c4862808424b3df6a1e27e7],Size_:374119253,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=844df08c-77a4-4ae3-83f6-20af3d43c259 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.449772134Z" level=info msg="Creating container: openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn/prometheus-operator-admission-webhook" id=a7a1e189-c17d-4d0d-9e3c-0b5f17a433a1 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.449919755Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:38.475341 2927 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-network-diagnostics/network-check-target-xmq2g] Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: I0105 09:29:37.873466 6409 ovs.go:90] Maximum command line arguments set to: 191102 Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: 2023-01-05T09:29:38Z [verbose] Add: openshift-network-diagnostics:network-check-target-xmq2g:f364a949-09d7-4cfb-83ff-e532b822a557:ovn-kubernetes(ovn-kubernetes):eth0 {"cniVersion":"0.4.0","interfaces":[{"name":"8a313b3accfad32","mac":"5a:c2:b3:e5:79:de"},{"name":"eth0","mac":"0a:58:0a:83:00:08","sandbox":"/var/run/netns/c226eea0-515f-4498-858c-3c6f29b4b2db"}],"ips":[{"version":"4","interface":1,"address":"10.131.0.8/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: I0105 09:29:38.454042 6393 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-network-diagnostics", Name:"network-check-target-xmq2g", UID:"f364a949-09d7-4cfb-83ff-e532b822a557", APIVersion:"v1", ResourceVersion:"90010", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.8/23] from ovn-kubernetes Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.502104557Z" level=info msg="Got pod network &{Name:network-check-target-xmq2g Namespace:openshift-network-diagnostics ID:8a313b3accfad3269666173b7b1c660df2741df597e595c3c47a5ca7eb07e228 UID:f364a949-09d7-4cfb-83ff-e532b822a557 NetNS:/var/run/netns/c226eea0-515f-4498-858c-3c6f29b4b2db Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.502309272Z" level=info msg="Checking pod openshift-network-diagnostics_network-check-target-xmq2g for CNI network multus-cni-network (type=multus)" Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.533207767Z" level=info msg="Created container ffd4b07638def8d9e366b9d82e7c3af366fb82361bbee1659a1998e20a8e5b84: openshift-monitoring/prometheus-k8s-0/init-config-reloader" id=0ab117f2-dbca-4820-a442-d3bfac49c381 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.632498372Z" level=info msg="Ran pod sandbox 8a313b3accfad3269666173b7b1c660df2741df597e595c3c47a5ca7eb07e228 with infra container: openshift-network-diagnostics/network-check-target-xmq2g/POD" id=b6c480b2-0246-48e3-b96d-279deb94c18e name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.643252447Z" level=info msg="Starting container: ffd4b07638def8d9e366b9d82e7c3af366fb82361bbee1659a1998e20a8e5b84" id=e4080c15-404e-49a9-972a-534168ce7df4 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.651497407Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7" id=b8fd73dc-8660-4954-b18b-40eee1c3e9ee name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.651705153Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:4cb41c73a232ec4401c5db32cdb523810dc12ac605cdc6a9ce388e90930eb6c4,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7],Size_:515955107,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b8fd73dc-8660-4954-b18b-40eee1c3e9ee name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.686688638Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7" id=c86c4698-153e-4270-bbca-0be837e423cf name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.687021179Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:4cb41c73a232ec4401c5db32cdb523810dc12ac605cdc6a9ce388e90930eb6c4,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7],Size_:515955107,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=c86c4698-153e-4270-bbca-0be837e423cf name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.687145247Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=6d8fbf15-c385-4238-9f9d-fbcc7bc99864 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.687328935Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=6d8fbf15-c385-4238-9f9d-fbcc7bc99864 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.693388746Z" level=info msg="Creating container: openshift-network-diagnostics/network-check-target-xmq2g/network-check-target-container" id=b8a5792e-f394-40d3-89a5-9ed995bac53e name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.693523146Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.695974701Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=6f5674ff-c316-4273-8f3b-a8fc16e3e0e8 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.696304014Z" level=info msg="Started container" PID=6532 containerID=dbf94c2e9e290b7a0ff09b82e88404bfcefb5c50d5c57120c11ad6bd75e3d27b description=openshift-multus/network-metrics-daemon-hq594/network-metrics-daemon id=7be066f8-5e0b-4f5b-997a-f3e6ab8d460f name=/runtime.v1.RuntimeService/StartContainer sandboxID=592c0fe68eaae38ffff4788882a53055cedb618ddc573956838720a91f262ddf Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:38.724392 2927 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-ingress/router-default-5cf6bbdff9-tnffs] Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.744474758Z" level=info msg="Started container" PID=6530 containerID=ffd4b07638def8d9e366b9d82e7c3af366fb82361bbee1659a1998e20a8e5b84 description=openshift-monitoring/prometheus-k8s-0/init-config-reloader id=e4080c15-404e-49a9-972a-534168ce7df4 name=/runtime.v1.RuntimeService/StartContainer sandboxID=04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0 Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: I0105 09:29:38.090117 6347 ovs.go:90] Maximum command line arguments set to: 191102 Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: 2023-01-05T09:29:38Z [verbose] Add: openshift-ingress:router-default-5cf6bbdff9-tnffs:c3b10ff4-a243-4a4e-b27d-f613e258666d:ovn-kubernetes(ovn-kubernetes):eth0 {"cniVersion":"0.4.0","interfaces":[{"name":"68d9bc9c6a9bcf2","mac":"fa:4b:fc:e1:56:a7"},{"name":"eth0","mac":"0a:58:0a:83:00:0a","sandbox":"/var/run/netns/b8f70c62-7b79-4f76-b3e5-1851b2785ffa"}],"ips":[{"version":"4","interface":1,"address":"10.131.0.10/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: I0105 09:29:38.688375 6304 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-ingress", Name:"router-default-5cf6bbdff9-tnffs", UID:"c3b10ff4-a243-4a4e-b27d-f613e258666d", APIVersion:"v1", ResourceVersion:"90066", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.10/23] from ovn-kubernetes Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.753834980Z" level=info msg="Got pod network &{Name:router-default-5cf6bbdff9-tnffs Namespace:openshift-ingress ID:68d9bc9c6a9bcf27e7bcf19bef0f4b3c6707eefbb0ba0e388f51a8e663a95ad7 UID:c3b10ff4-a243-4a4e-b27d-f613e258666d NetNS:/var/run/netns/b8f70c62-7b79-4f76-b3e5-1851b2785ffa Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.754045642Z" level=info msg="Checking pod openshift-ingress_router-default-5cf6bbdff9-tnffs for CNI network multus-cni-network (type=multus)" Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.845500655Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=c7a75815-a6a6-446f-9f2d-62452bf7c0c6 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.874179077Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=6f5674ff-c316-4273-8f3b-a8fc16e3e0e8 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.874405911Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=c7a75815-a6a6-446f-9f2d-62452bf7c0c6 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.918855575Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=27e8f952-4029-456a-b71a-c3e8012bc6da name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.919134329Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=27e8f952-4029-456a-b71a-c3e8012bc6da name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.919492269Z" level=info msg="Creating container: openshift-monitoring/telemeter-client-5598c645c4-mftdv/reload" id=2fff19b4-22f0-45b8-9c89-1882968248f4 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.919619258Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.935746219Z" level=info msg="Creating container: openshift-multus/network-metrics-daemon-hq594/kube-rbac-proxy" id=930030ed-bebb-4ff4-a770-aef2b39f3b3c name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.935865995Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:29:38.940078 2927 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3b10ff4_a243_4a4e_b27d_f613e258666d.slice/crio-68d9bc9c6a9bcf27e7bcf19bef0f4b3c6707eefbb0ba0e388f51a8e663a95ad7.scope WatchSource:0}: Error finding container 68d9bc9c6a9bcf27e7bcf19bef0f4b3c6707eefbb0ba0e388f51a8e663a95ad7: Status 404 returned error can't find the container with id 68d9bc9c6a9bcf27e7bcf19bef0f4b3c6707eefbb0ba0e388f51a8e663a95ad7 Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:38.941475 2927 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/kube-state-metrics-6798b548b9-xl88q] Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.959282997Z" level=info msg="Ran pod sandbox 68d9bc9c6a9bcf27e7bcf19bef0f4b3c6707eefbb0ba0e388f51a8e663a95ad7 with infra container: openshift-ingress/router-default-5cf6bbdff9-tnffs/POD" id=98a61853-3c1c-45d2-9d24-5548fa947bdc name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.963277734Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:bfd14188641a2d34e5b2b2ed3336b3b440155533f3fd7c6de776300dc3ff9e3b" id=e89f171c-5143-4e01-b52c-cf865d033072 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: I0105 09:29:38.382191 6475 ovs.go:90] Maximum command line arguments set to: 191102 Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: 2023-01-05T09:29:38Z [verbose] Add: openshift-monitoring:kube-state-metrics-6798b548b9-xl88q:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e:ovn-kubernetes(ovn-kubernetes):eth0 {"cniVersion":"0.4.0","interfaces":[{"name":"9c318dae338ceca","mac":"7a:a1:ce:94:dc:af"},{"name":"eth0","mac":"0a:58:0a:83:00:0d","sandbox":"/var/run/netns/8598084b-1708-4476-81ee-e77d49e075ea"}],"ips":[{"version":"4","interface":1,"address":"10.131.0.13/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: I0105 09:29:38.893477 6417 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"kube-state-metrics-6798b548b9-xl88q", UID:"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e", APIVersion:"v1", ResourceVersion:"90088", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.13/23] from ovn-kubernetes Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.964727533Z" level=info msg="Got pod network &{Name:kube-state-metrics-6798b548b9-xl88q Namespace:openshift-monitoring ID:9c318dae338ceca029f5b8d71eed7d2265d6188623995ec40e8d80da65a8d993 UID:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e NetNS:/var/run/netns/8598084b-1708-4476-81ee-e77d49e075ea Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:38.964872656Z" level=info msg="Checking pod openshift-monitoring_kube-state-metrics-6798b548b9-xl88q for CNI network multus-cni-network (type=multus)" Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:39.033253 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hq594" event=&{ID:50152296-9489-4fa8-aa42-7858debe1c08 Type:ContainerStarted Data:dbf94c2e9e290b7a0ff09b82e88404bfcefb5c50d5c57120c11ad6bd75e3d27b} Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.038445207Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:ad00ebc82f5cc671785b8401477b42db226d1c335ea60a4bb39bf28a961de840,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:bfd14188641a2d34e5b2b2ed3336b3b440155533f3fd7c6de776300dc3ff9e3b],Size_:411774965,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=e89f171c-5143-4e01-b52c-cf865d033072 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.048274331Z" level=info msg="Ran pod sandbox 9c318dae338ceca029f5b8d71eed7d2265d6188623995ec40e8d80da65a8d993 with infra container: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/POD" id=3b934532-6ee6-4d74-a5b1-22db4939f2d8 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.063468036Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:bfd14188641a2d34e5b2b2ed3336b3b440155533f3fd7c6de776300dc3ff9e3b" id=8f524fa6-aa8b-48a4-9475-c36bbe11e4f5 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.075855863Z" level=info msg="Created container 3d7374a8e3df4cc930990f545fee87038e1b2ad2278644240152875a9273e862: openshift-monitoring/alertmanager-main-0/alertmanager" id=f18bf7cd-9b19-407e-b0b2-c71686dccc94 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.082332152Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:ad00ebc82f5cc671785b8401477b42db226d1c335ea60a4bb39bf28a961de840,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:bfd14188641a2d34e5b2b2ed3336b3b440155533f3fd7c6de776300dc3ff9e3b],Size_:411774965,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=8f524fa6-aa8b-48a4-9475-c36bbe11e4f5 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.107294634Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e5521a3d1f64483e63ffe5bc23d9f882781cd4d152ac6de72013eaefe6dfcf89" id=d3b88c6f-d919-4324-a370-be04bd762a52 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.107500950Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:af30ce06a0837328332ff57ed27c48095a8ed9600535c919aa69fa76524919c6,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e5521a3d1f64483e63ffe5bc23d9f882781cd4d152ac6de72013eaefe6dfcf89],Size_:371510113,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=d3b88c6f-d919-4324-a370-be04bd762a52 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.125424769Z" level=info msg="Creating container: openshift-ingress/router-default-5cf6bbdff9-tnffs/router" id=f696123a-1bab-4056-8fe4-1369599b98a7 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.125538874Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.132142488Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e5521a3d1f64483e63ffe5bc23d9f882781cd4d152ac6de72013eaefe6dfcf89" id=b8d39f88-0f3e-4eb5-981e-a8427e75f5ea name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.132260488Z" level=info msg="Starting container: 3d7374a8e3df4cc930990f545fee87038e1b2ad2278644240152875a9273e862" id=2f57c1e7-d00e-4a46-a798-7d3ee1aea944 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:39.141210 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event=&{ID:05c9a6ea-140a-49a2-bde5-f220b05aa252 Type:ContainerStarted Data:de5e12e9744f461b9b6c0146346e5117bb31a1f6538299e5dccde6d8250c4268} Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:39.163371 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-8xrbh" event=&{ID:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748 Type:ContainerStarted Data:f226c6f82e17c96c33e4b6a3a2d81c74b2301377f9323d0a64195f4af1cc5679} Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.183628137Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:af30ce06a0837328332ff57ed27c48095a8ed9600535c919aa69fa76524919c6,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e5521a3d1f64483e63ffe5bc23d9f882781cd4d152ac6de72013eaefe6dfcf89],Size_:371510113,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b8d39f88-0f3e-4eb5-981e-a8427e75f5ea name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.251587124Z" level=info msg="Creating container: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-state-metrics" id=1a243380-d5d4-478e-99ab-a3e62ddb5eb8 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.251741678Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.269639059Z" level=info msg="Started container" PID=6682 containerID=3d7374a8e3df4cc930990f545fee87038e1b2ad2278644240152875a9273e862 description=openshift-monitoring/alertmanager-main-0/alertmanager id=2f57c1e7-d00e-4a46-a798-7d3ee1aea944 name=/runtime.v1.RuntimeService/StartContainer sandboxID=de5e12e9744f461b9b6c0146346e5117bb31a1f6538299e5dccde6d8250c4268 Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:39.294204 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" event=&{ID:a40bf20d-67a4-45e7-b701-ca5a75376c1f Type:ContainerStarted Data:fe3e7b123db46a0b88eddb5d106d43267fc1d0c1707470c0568f2b211ee6ba55} Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.306296069Z" level=info msg="Created container 75e3e62829960934d9fc0efc4aa383fbc6135b8a25d451f9e95cc3ccb510acad: openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh/prometheus-adapter" id=16a2b9b7-bb41-4c14-8921-d3b711789de4 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.310046002Z" level=info msg="Starting container: 75e3e62829960934d9fc0efc4aa383fbc6135b8a25d451f9e95cc3ccb510acad" id=0f6e74ce-bcf9-4766-8762-63cbb19e9ca8 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:39.319855 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xmq2g" event=&{ID:f364a949-09d7-4cfb-83ff-e532b822a557 Type:ContainerStarted Data:8a313b3accfad3269666173b7b1c660df2741df597e595c3c47a5ca7eb07e228} Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:39.359763 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" event=&{ID:88745ab7-efe7-42c0-a2e8-75591518333d Type:ContainerStarted Data:90d2b912f09210aef9996180fedf7e65296d392716fb68985112976a13c8d29b} Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.519408094Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=164d741d-8028-4314-92da-87159868b3ba name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.519638210Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=164d741d-8028-4314-92da-87159868b3ba name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.537581090Z" level=info msg="Started container" PID=6695 containerID=75e3e62829960934d9fc0efc4aa383fbc6135b8a25d451f9e95cc3ccb510acad description=openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh/prometheus-adapter id=0f6e74ce-bcf9-4766-8762-63cbb19e9ca8 name=/runtime.v1.RuntimeService/StartContainer sandboxID=fe3e7b123db46a0b88eddb5d106d43267fc1d0c1707470c0568f2b211ee6ba55 Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.550447404Z" level=info msg="Created container f26e0c4f3463c3fc98268d50d3fc2648f5e9d6728f5abbc2ae1088d62d5cca35: openshift-ingress-canary/ingress-canary-8xrbh/serve-healthcheck-canary" id=7a2a51dc-cd2b-4c0f-8027-8a084dcb96bd name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.550877833Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=f527dbd4-47c6-4969-9ee4-de6eed671e16 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.551124348Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=f527dbd4-47c6-4969-9ee4-de6eed671e16 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:39.547308 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" event=&{ID:fc45e847-8f55-4c48-8c48-4bfee2419dca Type:ContainerStarted Data:474834dd0b54abe011163e7458744c9ca3a52b894e6a07cfb86b009916a74367} Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:39.556136 2927 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-dns/dns-default-2pfzf] Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.556699018Z" level=info msg="Starting container: f26e0c4f3463c3fc98268d50d3fc2648f5e9d6728f5abbc2ae1088d62d5cca35" id=fb3a6077-1507-4e8b-a38b-067c40a5f48a name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.559472383Z" level=info msg="Creating container: openshift-monitoring/alertmanager-main-0/config-reloader" id=d717c0ec-c585-4d95-8a24-5abf10f7d0a1 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.559592732Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.602382412Z" level=info msg="Created container f81d81f334991a1a28d4bd7f1196fbd098494ebccd408af1173ec1fe9616bd86: openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn/prometheus-operator-admission-webhook" id=a7a1e189-c17d-4d0d-9e3c-0b5f17a433a1 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: I0105 09:29:38.892688 6521 ovs.go:90] Maximum command line arguments set to: 191102 Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: 2023-01-05T09:29:39Z [verbose] Add: openshift-dns:dns-default-2pfzf:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63:ovn-kubernetes(ovn-kubernetes):eth0 {"cniVersion":"0.4.0","interfaces":[{"name":"238eef320be8a6f","mac":"b2:55:05:aa:d1:4f"},{"name":"eth0","mac":"0a:58:0a:83:00:04","sandbox":"/var/run/netns/8e512d8a-b52a-45da-a484-9148509a8cea"}],"ips":[{"version":"4","interface":1,"address":"10.131.0.4/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: I0105 09:29:39.508694 6459 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-dns", Name:"dns-default-2pfzf", UID:"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63", APIVersion:"v1", ResourceVersion:"89963", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.4/23] from ovn-kubernetes Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.603059486Z" level=info msg="Got pod network &{Name:dns-default-2pfzf Namespace:openshift-dns ID:238eef320be8a6fa61ec83ea4bf9cdc139cde1247f51dbe0bd0040a8e9c70088 UID:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 NetNS:/var/run/netns/8e512d8a-b52a-45da-a484-9148509a8cea Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.603222891Z" level=info msg="Checking pod openshift-dns_dns-default-2pfzf for CNI network multus-cni-network (type=multus)" Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.635412718Z" level=info msg="Starting container: f81d81f334991a1a28d4bd7f1196fbd098494ebccd408af1173ec1fe9616bd86" id=cbf5e205-4e15-4170-9f8c-5d1b819b195b name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.713354999Z" level=info msg="Ran pod sandbox 238eef320be8a6fa61ec83ea4bf9cdc139cde1247f51dbe0bd0040a8e9c70088 with infra container: openshift-dns/dns-default-2pfzf/POD" id=349ce886-a7b4-4d61-8900-38fba3bfc31d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.733503978Z" level=info msg="Started container" PID=6746 containerID=f26e0c4f3463c3fc98268d50d3fc2648f5e9d6728f5abbc2ae1088d62d5cca35 description=openshift-ingress-canary/ingress-canary-8xrbh/serve-healthcheck-canary id=fb3a6077-1507-4e8b-a38b-067c40a5f48a name=/runtime.v1.RuntimeService/StartContainer sandboxID=f226c6f82e17c96c33e4b6a3a2d81c74b2301377f9323d0a64195f4af1cc5679 Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.765999541Z" level=info msg="Started container" PID=6752 containerID=f81d81f334991a1a28d4bd7f1196fbd098494ebccd408af1173ec1fe9616bd86 description=openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn/prometheus-operator-admission-webhook id=cbf5e205-4e15-4170-9f8c-5d1b819b195b name=/runtime.v1.RuntimeService/StartContainer sandboxID=90d2b912f09210aef9996180fedf7e65296d392716fb68985112976a13c8d29b Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.774180581Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f674b597694b13b7060c9ee611a403f1f87a98e05682c33f9a86140061ee9d9e" id=1640fed5-633d-4853-a095-ee476ff2cd8b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.774432795Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:cdb247afcbd9968f1a59680dcc4de0364f7ddd8fc25dd3a12db59f05c8e026a0,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f674b597694b13b7060c9ee611a403f1f87a98e05682c33f9a86140061ee9d9e],Size_:404212434,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=1640fed5-633d-4853-a095-ee476ff2cd8b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.786664876Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f674b597694b13b7060c9ee611a403f1f87a98e05682c33f9a86140061ee9d9e" id=0f3d0df7-88b1-4714-a0dd-3e5d99d46e0c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.786874253Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:cdb247afcbd9968f1a59680dcc4de0364f7ddd8fc25dd3a12db59f05c8e026a0,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f674b597694b13b7060c9ee611a403f1f87a98e05682c33f9a86140061ee9d9e],Size_:404212434,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=0f3d0df7-88b1-4714-a0dd-3e5d99d46e0c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.818103391Z" level=info msg="Creating container: openshift-dns/dns-default-2pfzf/dns" id=0f022951-ede5-4933-882a-fecd5e4fe100 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.818249839Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.948189743Z" level=info msg="Created container 81d748520620965f53fabe92f19a5bb3c5d4e2f876e096dcb82692d2f2f8f610: openshift-ingress/router-default-5cf6bbdff9-tnffs/router" id=f696123a-1bab-4056-8fe4-1369599b98a7 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:39.950519025Z" level=info msg="Starting container: 81d748520620965f53fabe92f19a5bb3c5d4e2f876e096dcb82692d2f2f8f610" id=5b553ef4-03a8-428c-8795-3afa50dc90cd name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:39.997940 2927 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256] Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: I0105 09:29:39.432404 6637 ovs.go:90] Maximum command line arguments set to: 191102 Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: 2023-01-05T09:29:39Z [verbose] Add: openshift-monitoring:openshift-state-metrics-6f5dcdf79b-wq256:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6:ovn-kubernetes(ovn-kubernetes):eth0 {"cniVersion":"0.4.0","interfaces":[{"name":"318b2d4d118085d","mac":"86:be:09:3c:1a:cb"},{"name":"eth0","mac":"0a:58:0a:83:00:05","sandbox":"/var/run/netns/d458dea1-0719-4a13-bc09-29a80e493860"}],"ips":[{"version":"4","interface":1,"address":"10.131.0.5/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: I0105 09:29:39.915471 6570 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"openshift-state-metrics-6f5dcdf79b-wq256", UID:"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6", APIVersion:"v1", ResourceVersion:"89967", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.5/23] from ovn-kubernetes Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.030255274Z" level=info msg="Got pod network &{Name:openshift-state-metrics-6f5dcdf79b-wq256 Namespace:openshift-monitoring ID:318b2d4d118085d70f305dfbfb7875ec65ba1a3f9539d6e1f29f152cb2d8af82 UID:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 NetNS:/var/run/netns/d458dea1-0719-4a13-bc09-29a80e493860 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.030425055Z" level=info msg="Checking pod openshift-monitoring_openshift-state-metrics-6f5dcdf79b-wq256 for CNI network multus-cni-network (type=multus)" Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.040133135Z" level=info msg="Ran pod sandbox 318b2d4d118085d70f305dfbfb7875ec65ba1a3f9539d6e1f29f152cb2d8af82 with infra container: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/POD" id=fc9103cb-ce68-45a9-b08a-00886595b7e5 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.041234813Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=04b6d891-fa4d-41b7-b737-bf7bf690959e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.041514495Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=04b6d891-fa4d-41b7-b737-bf7bf690959e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.042632544Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=6fb95f42-8dfb-4f13-ac7d-add8212b38a0 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.042850730Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=6fb95f42-8dfb-4f13-ac7d-add8212b38a0 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.043484050Z" level=info msg="Started container" PID=6883 containerID=81d748520620965f53fabe92f19a5bb3c5d4e2f876e096dcb82692d2f2f8f610 description=openshift-ingress/router-default-5cf6bbdff9-tnffs/router id=5b553ef4-03a8-428c-8795-3afa50dc90cd name=/runtime.v1.RuntimeService/StartContainer sandboxID=68d9bc9c6a9bcf27e7bcf19bef0f4b3c6707eefbb0ba0e388f51a8e663a95ad7 Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.044047563Z" level=info msg="Creating container: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/kube-rbac-proxy-main" id=5ff2e2aa-6aba-4119-8fd6-1e123d5ddcd4 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.044154473Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.120184651Z" level=info msg="Created container 1a6952a00cea2c0e412107ad3a67144e3d305ab025013ce4a8271b69e5bf514c: openshift-network-diagnostics/network-check-target-xmq2g/network-check-target-container" id=b8a5792e-f394-40d3-89a5-9ed995bac53e name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.129719754Z" level=info msg="Created container 65d2017d651f03a83c1df13113775d03f51374c0b6714011dd6c4783e504bbc7: openshift-monitoring/alertmanager-main-0/config-reloader" id=d717c0ec-c585-4d95-8a24-5abf10f7d0a1 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.134120370Z" level=info msg="Starting container: 1a6952a00cea2c0e412107ad3a67144e3d305ab025013ce4a8271b69e5bf514c" id=1541347b-da82-497d-a37b-bdea7bfdb9b8 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.156836953Z" level=info msg="Starting container: 65d2017d651f03a83c1df13113775d03f51374c0b6714011dd6c4783e504bbc7" id=ad15687f-9d4c-4699-9636-fc0cba4cf43f name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.172298184Z" level=info msg="Created container a04d63e53a7fcdfd5ab131841815a2f2677c7f07a565fb1c56e7fe2739ce6414: openshift-monitoring/telemeter-client-5598c645c4-mftdv/reload" id=2fff19b4-22f0-45b8-9c89-1882968248f4 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.180926534Z" level=info msg="Starting container: a04d63e53a7fcdfd5ab131841815a2f2677c7f07a565fb1c56e7fe2739ce6414" id=2cab84e2-f33f-485b-bad9-ed7c142c56e1 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.238416496Z" level=info msg="Started container" PID=6926 containerID=65d2017d651f03a83c1df13113775d03f51374c0b6714011dd6c4783e504bbc7 description=openshift-monitoring/alertmanager-main-0/config-reloader id=ad15687f-9d4c-4699-9636-fc0cba4cf43f name=/runtime.v1.RuntimeService/StartContainer sandboxID=de5e12e9744f461b9b6c0146346e5117bb31a1f6538299e5dccde6d8250c4268 Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.238588996Z" level=info msg="Started container" PID=6867 containerID=1a6952a00cea2c0e412107ad3a67144e3d305ab025013ce4a8271b69e5bf514c description=openshift-network-diagnostics/network-check-target-xmq2g/network-check-target-container id=1541347b-da82-497d-a37b-bdea7bfdb9b8 name=/runtime.v1.RuntimeService/StartContainer sandboxID=8a313b3accfad3269666173b7b1c660df2741df597e595c3c47a5ca7eb07e228 Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.273621767Z" level=info msg="Started container" PID=6870 containerID=a04d63e53a7fcdfd5ab131841815a2f2677c7f07a565fb1c56e7fe2739ce6414 description=openshift-monitoring/telemeter-client-5598c645c4-mftdv/reload id=2cab84e2-f33f-485b-bad9-ed7c142c56e1 name=/runtime.v1.RuntimeService/StartContainer sandboxID=f1e02d518b5e62edc187ed7cc730fab2279f2eb5384e3f3ffd25d3a014b31d4b Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.297304044Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=b969cd41-46bc-4a66-9ac5-ac63daa8cbaf name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.297567400Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b969cd41-46bc-4a66-9ac5-ac63daa8cbaf name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.299598820Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=f153cf62-ff13-4163-8329-b603d1ed314b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.299786456Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=f153cf62-ff13-4163-8329-b603d1ed314b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.301718991Z" level=info msg="Creating container: openshift-monitoring/alertmanager-main-0/alertmanager-proxy" id=884a121e-86fe-4ad4-8295-d0d37f9273d8 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.301839442Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.320101883Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=2a6220a6-75d8-4395-bf1b-2999a91e2063 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.342377389Z" level=info msg="Created container 9c70e69ff06eb3aacb6dd1829475cbd0126780b7fa69a099ad7ca98bac92ddef: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-state-metrics" id=1a243380-d5d4-478e-99ab-a3e62ddb5eb8 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.342967587Z" level=info msg="Starting container: 9c70e69ff06eb3aacb6dd1829475cbd0126780b7fa69a099ad7ca98bac92ddef" id=fdac597d-0b3f-49c9-9c7e-0a85d37ca65a name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.355147322Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=2a6220a6-75d8-4395-bf1b-2999a91e2063 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.366367423Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=0c1b9905-e3c0-4553-bcf3-4a08623fa823 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.366562465Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=0c1b9905-e3c0-4553-bcf3-4a08623fa823 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.398748472Z" level=info msg="Creating container: openshift-monitoring/telemeter-client-5598c645c4-mftdv/kube-rbac-proxy" id=f8da39ce-8c23-47e7-b65d-2baf752f0a0d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.398993217Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.413234871Z" level=info msg="Started container" PID=6946 containerID=9c70e69ff06eb3aacb6dd1829475cbd0126780b7fa69a099ad7ca98bac92ddef description=openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-state-metrics id=fdac597d-0b3f-49c9-9c7e-0a85d37ca65a name=/runtime.v1.RuntimeService/StartContainer sandboxID=9c318dae338ceca029f5b8d71eed7d2265d6188623995ec40e8d80da65a8d993 Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.421103267Z" level=info msg="Created container 6472505d47cca16f78859f90a7fe51edb203e622be56c16aff7c6021f648623f: openshift-multus/network-metrics-daemon-hq594/kube-rbac-proxy" id=930030ed-bebb-4ff4-a770-aef2b39f3b3c name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.422787968Z" level=info msg="Starting container: 6472505d47cca16f78859f90a7fe51edb203e622be56c16aff7c6021f648623f" id=46a047ea-b408-46b5-881c-5d2fa9497f23 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.529209323Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=27ea9c2c-bb22-4aaa-9e8e-1ed34c2231bd name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.529440607Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=27ea9c2c-bb22-4aaa-9e8e-1ed34c2231bd name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.530484357Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=ebf8d3f5-4c36-4542-a251-a718bab7e5b0 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.530678484Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=ebf8d3f5-4c36-4542-a251-a718bab7e5b0 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.532122078Z" level=info msg="Creating container: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-rbac-proxy-main" id=a927fce0-07f7-43b3-ae34-be1a6173080e name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.532238865Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.555590315Z" level=info msg="Started container" PID=7003 containerID=6472505d47cca16f78859f90a7fe51edb203e622be56c16aff7c6021f648623f description=openshift-multus/network-metrics-daemon-hq594/kube-rbac-proxy id=46a047ea-b408-46b5-881c-5d2fa9497f23 name=/runtime.v1.RuntimeService/StartContainer sandboxID=592c0fe68eaae38ffff4788882a53055cedb618ddc573956838720a91f262ddf Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:40.582861 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event=&{ID:05c9a6ea-140a-49a2-bde5-f220b05aa252 Type:ContainerStarted Data:65d2017d651f03a83c1df13113775d03f51374c0b6714011dd6c4783e504bbc7} Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:40.582929 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event=&{ID:05c9a6ea-140a-49a2-bde5-f220b05aa252 Type:ContainerStarted Data:3d7374a8e3df4cc930990f545fee87038e1b2ad2278644240152875a9273e862} Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:40.583926 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-8xrbh" event=&{ID:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748 Type:ContainerStarted Data:f26e0c4f3463c3fc98268d50d3fc2648f5e9d6728f5abbc2ae1088d62d5cca35} Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:40.585676 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" event=&{ID:a40bf20d-67a4-45e7-b701-ca5a75376c1f Type:ContainerStarted Data:75e3e62829960934d9fc0efc4aa383fbc6135b8a25d451f9e95cc3ccb510acad} Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:40.586670 2927 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:40.659730 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xmq2g" event=&{ID:f364a949-09d7-4cfb-83ff-e532b822a557 Type:ContainerStarted Data:1a6952a00cea2c0e412107ad3a67144e3d305ab025013ce4a8271b69e5bf514c} Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:40.660416 2927 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:40.661494 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" event=&{ID:88745ab7-efe7-42c0-a2e8-75591518333d Type:ContainerStarted Data:f81d81f334991a1a28d4bd7f1196fbd098494ebccd408af1173ec1fe9616bd86} Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:40.662424 2927 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:40.674925 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" event=&{ID:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 Type:ContainerStarted Data:318b2d4d118085d70f305dfbfb7875ec65ba1a3f9539d6e1f29f152cb2d8af82} Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:40.729223 2927 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:40.730587 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" event=&{ID:fc45e847-8f55-4c48-8c48-4bfee2419dca Type:ContainerStarted Data:a04d63e53a7fcdfd5ab131841815a2f2677c7f07a565fb1c56e7fe2739ce6414} Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:40.732045 2927 generic.go:296] "Generic (PLEG): container finished" podID=4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 containerID="ffd4b07638def8d9e366b9d82e7c3af366fb82361bbee1659a1998e20a8e5b84" exitCode=0 Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:40.732101 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event=&{ID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Type:ContainerDied Data:ffd4b07638def8d9e366b9d82e7c3af366fb82361bbee1659a1998e20a8e5b84} Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.741795207Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:91e7d68b28aa00e209065604b2b9ac1400671d9102581e8b49b37726acd7b6e4" id=b9a5bcf3-f844-457b-a5b5-1da3ac091c06 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.742180794Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b68b0fa7ea4653b56a91a8c213d456cab2f645207af609598f9877a41585eeae,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:91e7d68b28aa00e209065604b2b9ac1400671d9102581e8b49b37726acd7b6e4],Size_:493220384,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b9a5bcf3-f844-457b-a5b5-1da3ac091c06 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:40.756457 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" event=&{ID:c3b10ff4-a243-4a4e-b27d-f613e258666d Type:ContainerStarted Data:81d748520620965f53fabe92f19a5bb3c5d4e2f876e096dcb82692d2f2f8f610} Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:40.756487 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" event=&{ID:c3b10ff4-a243-4a4e-b27d-f613e258666d Type:ContainerStarted Data:68d9bc9c6a9bcf27e7bcf19bef0f4b3c6707eefbb0ba0e388f51a8e663a95ad7} Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.784721563Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:91e7d68b28aa00e209065604b2b9ac1400671d9102581e8b49b37726acd7b6e4" id=bb04efcd-5246-4572-a737-758a201defd8 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.784991124Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b68b0fa7ea4653b56a91a8c213d456cab2f645207af609598f9877a41585eeae,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:91e7d68b28aa00e209065604b2b9ac1400671d9102581e8b49b37726acd7b6e4],Size_:493220384,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=bb04efcd-5246-4572-a737-758a201defd8 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.787853271Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-0/prometheus" id=eacea3fb-4596-49b0-9c34-042b401e0011 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.788004422Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:40.802192 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" event=&{ID:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e Type:ContainerStarted Data:9c70e69ff06eb3aacb6dd1829475cbd0126780b7fa69a099ad7ca98bac92ddef} Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:40.802232 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" event=&{ID:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e Type:ContainerStarted Data:9c318dae338ceca029f5b8d71eed7d2265d6188623995ec40e8d80da65a8d993} Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:40.819960 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-2pfzf" event=&{ID:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 Type:ContainerStarted Data:238eef320be8a6fa61ec83ea4bf9cdc139cde1247f51dbe0bd0040a8e9c70088} Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.821049538Z" level=info msg="Created container 011acc34c572433812691a69859a1ec4d3706557a1b172add408479127957e6d: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/kube-rbac-proxy-main" id=5ff2e2aa-6aba-4119-8fd6-1e123d5ddcd4 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.822804825Z" level=info msg="Starting container: 011acc34c572433812691a69859a1ec4d3706557a1b172add408479127957e6d" id=79b17f6b-06a0-46e6-955f-8c4f4459b12c name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.918569653Z" level=info msg="Created container 1c1ff2a5effcdc52ca7d965b85b5b36131e77e34ca25d1e8807eedb4eb9c99d0: openshift-monitoring/alertmanager-main-0/alertmanager-proxy" id=884a121e-86fe-4ad4-8295-d0d37f9273d8 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.926607909Z" level=info msg="Started container" PID=7067 containerID=011acc34c572433812691a69859a1ec4d3706557a1b172add408479127957e6d description=openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/kube-rbac-proxy-main id=79b17f6b-06a0-46e6-955f-8c4f4459b12c name=/runtime.v1.RuntimeService/StartContainer sandboxID=318b2d4d118085d70f305dfbfb7875ec65ba1a3f9539d6e1f29f152cb2d8af82 Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.926788780Z" level=info msg="Starting container: 1c1ff2a5effcdc52ca7d965b85b5b36131e77e34ca25d1e8807eedb4eb9c99d0" id=2bd74cb7-a2ce-4eef-b5c2-7607c2906d07 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.978704751Z" level=info msg="Started container" PID=7117 containerID=1c1ff2a5effcdc52ca7d965b85b5b36131e77e34ca25d1e8807eedb4eb9c99d0 description=openshift-monitoring/alertmanager-main-0/alertmanager-proxy id=2bd74cb7-a2ce-4eef-b5c2-7607c2906d07 name=/runtime.v1.RuntimeService/StartContainer sandboxID=de5e12e9744f461b9b6c0146346e5117bb31a1f6538299e5dccde6d8250c4268 Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.982427019Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=9aa16033-d44b-4c88-a2f0-5bd141e579c1 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:40.982643801Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=9aa16033-d44b-4c88-a2f0-5bd141e579c1 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.000772783Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=ee7c8ad6-efa2-4d51-8f5f-01fc81fb97fe name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.001071266Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=ee7c8ad6-efa2-4d51-8f5f-01fc81fb97fe name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.002182428Z" level=info msg="Creating container: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/kube-rbac-proxy-self" id=353b1295-2598-4398-b7fb-533af502819f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.002300889Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.038937833Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=92c4f74e-c626-42a0-941f-2e9320b73c5f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.069678237Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=92c4f74e-c626-42a0-941f-2e9320b73c5f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.073946096Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=5db0665c-a258-4754-8cb2-13494ce262a5 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.074150582Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=5db0665c-a258-4754-8cb2-13494ce262a5 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.075163472Z" level=info msg="Creating container: openshift-monitoring/alertmanager-main-0/kube-rbac-proxy" id=09c303fa-2baa-49fd-a70c-dd228fa5990a name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.075269654Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:41.162364 2927 kubelet.go:2229] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:41.185205 2927 patch_prober.go:29] interesting pod/router-default-5cf6bbdff9-tnffs container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: [-]has-synced failed: reason withheld Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: [+]process-running ok Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: healthz check failed Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:41.185262 2927 prober.go:114] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" podUID=c3b10ff4-a243-4a4e-b27d-f613e258666d containerName="router" probeResult=failure output="HTTP probe failed with statuscode: 500" Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.243950463Z" level=info msg="Created container 3b77a73a551754a882d136a50cbe7d425606b96a3c681a87e9ec959b8bfe08ce: openshift-monitoring/telemeter-client-5598c645c4-mftdv/kube-rbac-proxy" id=f8da39ce-8c23-47e7-b65d-2baf752f0a0d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.245551297Z" level=info msg="Starting container: 3b77a73a551754a882d136a50cbe7d425606b96a3c681a87e9ec959b8bfe08ce" id=8570c10b-ee26-4990-806b-bb3b0d222a3a name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.273644704Z" level=info msg="Started container" PID=7168 containerID=3b77a73a551754a882d136a50cbe7d425606b96a3c681a87e9ec959b8bfe08ce description=openshift-monitoring/telemeter-client-5598c645c4-mftdv/kube-rbac-proxy id=8570c10b-ee26-4990-806b-bb3b0d222a3a name=/runtime.v1.RuntimeService/StartContainer sandboxID=f1e02d518b5e62edc187ed7cc730fab2279f2eb5384e3f3ffd25d3a014b31d4b Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.392132550Z" level=info msg="Created container 6e7bca9816716c79212a27fedd6718c5b95084ab803f37080afc6742d7525dc8: openshift-dns/dns-default-2pfzf/dns" id=0f022951-ede5-4933-882a-fecd5e4fe100 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.392702409Z" level=info msg="Starting container: 6e7bca9816716c79212a27fedd6718c5b95084ab803f37080afc6742d7525dc8" id=e8f8aebe-a435-4606-aab3-b94ba1d6a4c7 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.443200719Z" level=info msg="Started container" PID=7260 containerID=6e7bca9816716c79212a27fedd6718c5b95084ab803f37080afc6742d7525dc8 description=openshift-dns/dns-default-2pfzf/dns id=e8f8aebe-a435-4606-aab3-b94ba1d6a4c7 name=/runtime.v1.RuntimeService/StartContainer sandboxID=238eef320be8a6fa61ec83ea4bf9cdc139cde1247f51dbe0bd0040a8e9c70088 Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.460592889Z" level=info msg="Created container 876bc02802326ac406bee168382dda6d347e8d7eb8b3e236476da5e28e853f80: openshift-monitoring/prometheus-k8s-0/prometheus" id=eacea3fb-4596-49b0-9c34-042b401e0011 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.463998992Z" level=info msg="Starting container: 876bc02802326ac406bee168382dda6d347e8d7eb8b3e236476da5e28e853f80" id=a5ce2748-e273-40cb-a3db-76ce006174f8 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.475225438Z" level=info msg="Created container af38b04d912e1eb5b634a533910f05eb846127a534a3ac450ce9b629df1b1f1f: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-rbac-proxy-main" id=a927fce0-07f7-43b3-ae34-be1a6173080e name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.495261033Z" level=info msg="Starting container: af38b04d912e1eb5b634a533910f05eb846127a534a3ac450ce9b629df1b1f1f" id=2b6c11f9-cf18-48bc-8cfa-ca01f28f5e4d name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.538857142Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=2eeeeaf0-b9c3-42d2-874c-11bc71aca4a9 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.539119514Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=2eeeeaf0-b9c3-42d2-874c-11bc71aca4a9 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.540303909Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=891c83fe-7451-476e-b9d6-4db6ed47e895 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.540482484Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=891c83fe-7451-476e-b9d6-4db6ed47e895 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.541438282Z" level=info msg="Creating container: openshift-dns/dns-default-2pfzf/kube-rbac-proxy" id=80e5064a-8fa2-4a20-a0ca-51d4964b785d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.541553458Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.553715260Z" level=info msg="Started container" PID=7253 containerID=876bc02802326ac406bee168382dda6d347e8d7eb8b3e236476da5e28e853f80 description=openshift-monitoring/prometheus-k8s-0/prometheus id=a5ce2748-e273-40cb-a3db-76ce006174f8 name=/runtime.v1.RuntimeService/StartContainer sandboxID=04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0 Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.597486086Z" level=info msg="Started container" PID=7222 containerID=af38b04d912e1eb5b634a533910f05eb846127a534a3ac450ce9b629df1b1f1f description=openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-rbac-proxy-main id=2b6c11f9-cf18-48bc-8cfa-ca01f28f5e4d name=/runtime.v1.RuntimeService/StartContainer sandboxID=9c318dae338ceca029f5b8d71eed7d2265d6188623995ec40e8d80da65a8d993 Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.609563696Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=63781712-a62d-4143-8a0f-0dfd99e7ce22 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.631829700Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=63781712-a62d-4143-8a0f-0dfd99e7ce22 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.632772148Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=a1664a1b-d8eb-4d14-9a3e-15a70a2efa72 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.633028248Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=a1664a1b-d8eb-4d14-9a3e-15a70a2efa72 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.634007225Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-0/config-reloader" id=e85b753e-b9a9-49b8-b4e5-e1847939f471 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.634129601Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.636986782Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=8d31f95e-30f9-4ff0-a17d-a65cf4d00b0b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.637167268Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=8d31f95e-30f9-4ff0-a17d-a65cf4d00b0b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.637954121Z" level=info msg="Created container 1d6d515b36d97644368b6561572ffcff9d150ef935eef5e962c5fabb98e5369a: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/kube-rbac-proxy-self" id=353b1295-2598-4398-b7fb-533af502819f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.637972472Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=12b0c725-3e8a-4302-8a01-652ded3c00f9 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.638195931Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=12b0c725-3e8a-4302-8a01-652ded3c00f9 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.639246770Z" level=info msg="Creating container: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-rbac-proxy-self" id=10cc64f8-a9ab-4562-ae58-00c3c0a2405c name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.639349807Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.639731964Z" level=info msg="Starting container: 1d6d515b36d97644368b6561572ffcff9d150ef935eef5e962c5fabb98e5369a" id=72111624-6965-49ac-bafe-2c739138dfee name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.724293035Z" level=info msg="Started container" PID=7287 containerID=1d6d515b36d97644368b6561572ffcff9d150ef935eef5e962c5fabb98e5369a description=openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/kube-rbac-proxy-self id=72111624-6965-49ac-bafe-2c739138dfee name=/runtime.v1.RuntimeService/StartContainer sandboxID=318b2d4d118085d70f305dfbfb7875ec65ba1a3f9539d6e1f29f152cb2d8af82 Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.853438660Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:78252776c28ff0ef5195173aebd633dc3a71e5d4732ca01f9104bb0aab07570e" id=1efae2e8-a503-4106-abd4-ebc4d5bb3f27 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.879036314Z" level=info msg="Created container fbba8d600bd026ca0a23b1c9724088c18597af8b742d44a66208c8f6b0dbe0ad: openshift-monitoring/alertmanager-main-0/kube-rbac-proxy" id=09c303fa-2baa-49fd-a70c-dd228fa5990a name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.887192684Z" level=info msg="Starting container: fbba8d600bd026ca0a23b1c9724088c18597af8b742d44a66208c8f6b0dbe0ad" id=a70f9e9b-6ad4-4301-82a4-870e8b6bbb36 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:41.887877 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-2pfzf" event=&{ID:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 Type:ContainerStarted Data:6e7bca9816716c79212a27fedd6718c5b95084ab803f37080afc6742d7525dc8} Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:41.889348 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" event=&{ID:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e Type:ContainerStarted Data:af38b04d912e1eb5b634a533910f05eb846127a534a3ac450ce9b629df1b1f1f} Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:41.891291 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" event=&{ID:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 Type:ContainerStarted Data:1d6d515b36d97644368b6561572ffcff9d150ef935eef5e962c5fabb98e5369a} Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:41.891319 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" event=&{ID:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 Type:ContainerStarted Data:011acc34c572433812691a69859a1ec4d3706557a1b172add408479127957e6d} Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:41.893116 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" event=&{ID:fc45e847-8f55-4c48-8c48-4bfee2419dca Type:ContainerStarted Data:3b77a73a551754a882d136a50cbe7d425606b96a3c681a87e9ec959b8bfe08ce} Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:41.897696 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event=&{ID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Type:ContainerStarted Data:876bc02802326ac406bee168382dda6d347e8d7eb8b3e236476da5e28e853f80} Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:41.900510 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hq594" event=&{ID:50152296-9489-4fa8-aa42-7858debe1c08 Type:ContainerStarted Data:6472505d47cca16f78859f90a7fe51edb203e622be56c16aff7c6021f648623f} Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:41.903945 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event=&{ID:05c9a6ea-140a-49a2-bde5-f220b05aa252 Type:ContainerStarted Data:1c1ff2a5effcdc52ca7d965b85b5b36131e77e34ca25d1e8807eedb4eb9c99d0} Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.947730230Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:ab19b64ab85c6b98c0381b09c3bd6f6db37bb31b25f12aa07bf856e8be07eac7,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:78252776c28ff0ef5195173aebd633dc3a71e5d4732ca01f9104bb0aab07570e],Size_:365012751,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=1efae2e8-a503-4106-abd4-ebc4d5bb3f27 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.965255157Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:78252776c28ff0ef5195173aebd633dc3a71e5d4732ca01f9104bb0aab07570e" id=bdfed7ef-d0d2-41f4-aa14-c0bf0ee48eae name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.965487536Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:ab19b64ab85c6b98c0381b09c3bd6f6db37bb31b25f12aa07bf856e8be07eac7,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:78252776c28ff0ef5195173aebd633dc3a71e5d4732ca01f9104bb0aab07570e],Size_:365012751,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=bdfed7ef-d0d2-41f4-aa14-c0bf0ee48eae name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.966852021Z" level=info msg="Started container" PID=7313 containerID=fbba8d600bd026ca0a23b1c9724088c18597af8b742d44a66208c8f6b0dbe0ad description=openshift-monitoring/alertmanager-main-0/kube-rbac-proxy id=a70f9e9b-6ad4-4301-82a4-870e8b6bbb36 name=/runtime.v1.RuntimeService/StartContainer sandboxID=de5e12e9744f461b9b6c0146346e5117bb31a1f6538299e5dccde6d8250c4268 Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.972437154Z" level=info msg="Creating container: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/openshift-state-metrics" id=51db08aa-0129-47b8-ab3f-1124ac51425c name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:41.972561415Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.051213868Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=2c81319d-1398-4871-97f9-79b35646abeb name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.067329872Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=2c81319d-1398-4871-97f9-79b35646abeb name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.092537650Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=93d9c1ab-42f6-43c4-a216-b940ff599ee8 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.092759959Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=93d9c1ab-42f6-43c4-a216-b940ff599ee8 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.095046020Z" level=info msg="Creating container: openshift-monitoring/alertmanager-main-0/kube-rbac-proxy-metric" id=00c0f622-add0-4dae-be43-9fbe2495109f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.095151334Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:42.163476 2927 patch_prober.go:29] interesting pod/router-default-5cf6bbdff9-tnffs container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: [-]has-synced failed: reason withheld Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: [+]process-running ok Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: healthz check failed Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:42.163531 2927 prober.go:114] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" podUID=c3b10ff4-a243-4a4e-b27d-f613e258666d containerName="router" probeResult=failure output="HTTP probe failed with statuscode: 500" Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.194638716Z" level=info msg="Created container f218c2963bb329ec34cc051d3d32e91b0f36d73cbc9ea90fa4cdef808dbf1e48: openshift-dns/dns-default-2pfzf/kube-rbac-proxy" id=80e5064a-8fa2-4a20-a0ca-51d4964b785d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.196558242Z" level=info msg="Starting container: f218c2963bb329ec34cc051d3d32e91b0f36d73cbc9ea90fa4cdef808dbf1e48" id=1815ae24-8e56-4607-b931-84e560b7be16 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.217208302Z" level=info msg="Started container" PID=7439 containerID=f218c2963bb329ec34cc051d3d32e91b0f36d73cbc9ea90fa4cdef808dbf1e48 description=openshift-dns/dns-default-2pfzf/kube-rbac-proxy id=1815ae24-8e56-4607-b931-84e560b7be16 name=/runtime.v1.RuntimeService/StartContainer sandboxID=238eef320be8a6fa61ec83ea4bf9cdc139cde1247f51dbe0bd0040a8e9c70088 Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.563193869Z" level=info msg="Created container 3676f1603de80edd23f34248eef289ffee599487f3f97ddd83f3362b1b7dc2f0: openshift-monitoring/prometheus-k8s-0/config-reloader" id=e85b753e-b9a9-49b8-b4e5-e1847939f471 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.563332565Z" level=info msg="Created container a9fdb00607de8750f6e02b9aad69c1eb2f10b013ec69caa09bea48d801e3775d: openshift-monitoring/alertmanager-main-0/kube-rbac-proxy-metric" id=00c0f622-add0-4dae-be43-9fbe2495109f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.564010775Z" level=info msg="Starting container: 3676f1603de80edd23f34248eef289ffee599487f3f97ddd83f3362b1b7dc2f0" id=6e4e996c-948c-4ea1-8d9e-22f3f173e8ee name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.564146166Z" level=info msg="Starting container: a9fdb00607de8750f6e02b9aad69c1eb2f10b013ec69caa09bea48d801e3775d" id=4249def1-fab6-4c83-8010-6cf8d17bbd18 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.587126475Z" level=info msg="Created container b1800c468fe04ef1ee6fc5ecf08f9b47946f276d072facfdc0f9944871c84001: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/openshift-state-metrics" id=51db08aa-0129-47b8-ab3f-1124ac51425c name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.590817868Z" level=info msg="Starting container: b1800c468fe04ef1ee6fc5ecf08f9b47946f276d072facfdc0f9944871c84001" id=538be570-450a-440d-a40b-2eb2e027f244 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.616878814Z" level=info msg="Created container 047cddfaab1d98f2f4be03b086b5acc263c7285299eed9b70a12b9c6485a6435: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-rbac-proxy-self" id=10cc64f8-a9ab-4562-ae58-00c3c0a2405c name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.617718144Z" level=info msg="Starting container: 047cddfaab1d98f2f4be03b086b5acc263c7285299eed9b70a12b9c6485a6435" id=8bbebd2d-3952-4f2a-bc4d-15d120e9e41c name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.633797927Z" level=info msg="Started container" PID=7530 containerID=a9fdb00607de8750f6e02b9aad69c1eb2f10b013ec69caa09bea48d801e3775d description=openshift-monitoring/alertmanager-main-0/kube-rbac-proxy-metric id=4249def1-fab6-4c83-8010-6cf8d17bbd18 name=/runtime.v1.RuntimeService/StartContainer sandboxID=de5e12e9744f461b9b6c0146346e5117bb31a1f6538299e5dccde6d8250c4268 Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.665209154Z" level=info msg="Started container" PID=7494 containerID=3676f1603de80edd23f34248eef289ffee599487f3f97ddd83f3362b1b7dc2f0 description=openshift-monitoring/prometheus-k8s-0/config-reloader id=6e4e996c-948c-4ea1-8d9e-22f3f173e8ee name=/runtime.v1.RuntimeService/StartContainer sandboxID=04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0 Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.670285540Z" level=info msg="Started container" PID=7489 containerID=047cddfaab1d98f2f4be03b086b5acc263c7285299eed9b70a12b9c6485a6435 description=openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-rbac-proxy-self id=8bbebd2d-3952-4f2a-bc4d-15d120e9e41c name=/runtime.v1.RuntimeService/StartContainer sandboxID=9c318dae338ceca029f5b8d71eed7d2265d6188623995ec40e8d80da65a8d993 Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.670578603Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816" id=a6206d39-83ab-44ab-9136-f2057776c9f7 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.670823429Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:955ee2b41ecd88e3a9c0183a0b65d6edbf524cefa1318837dcbc82b0466c4d55,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816],Size_:309576938,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=a6206d39-83ab-44ab-9136-f2057776c9f7 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.672312792Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816" id=41cf7ce2-a4d4-472a-bcef-4085fda0127f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.672509403Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:955ee2b41ecd88e3a9c0183a0b65d6edbf524cefa1318837dcbc82b0466c4d55,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816],Size_:309576938,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=41cf7ce2-a4d4-472a-bcef-4085fda0127f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.673445519Z" level=info msg="Creating container: openshift-monitoring/alertmanager-main-0/prom-label-proxy" id=290cf590-b704-4f51-8012-cc48cb78b694 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.673569573Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.680878246Z" level=info msg="Started container" PID=7507 containerID=b1800c468fe04ef1ee6fc5ecf08f9b47946f276d072facfdc0f9944871c84001 description=openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/openshift-state-metrics id=538be570-450a-440d-a40b-2eb2e027f244 name=/runtime.v1.RuntimeService/StartContainer sandboxID=318b2d4d118085d70f305dfbfb7875ec65ba1a3f9539d6e1f29f152cb2d8af82 Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.737042829Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc" id=748deadd-dd91-4a14-b2b3-ac788e34f80e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.739922842Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:08ff7a43d9185529520a29e114f0c28b070000b464a44b347240015f3b466488,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc],Size_:368278089,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=748deadd-dd91-4a14-b2b3-ac788e34f80e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.755126478Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc" id=889dd8a0-8e5a-463b-bc4b-69fecac93ee8 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.755341152Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:08ff7a43d9185529520a29e114f0c28b070000b464a44b347240015f3b466488,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc],Size_:368278089,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=889dd8a0-8e5a-463b-bc4b-69fecac93ee8 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.772117474Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-0/thanos-sidecar" id=5a44d88d-4bcf-4e20-ad44-59302e23377e name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:42.772243114Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:42.910764 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event=&{ID:05c9a6ea-140a-49a2-bde5-f220b05aa252 Type:ContainerStarted Data:a9fdb00607de8750f6e02b9aad69c1eb2f10b013ec69caa09bea48d801e3775d} Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:42.910801 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event=&{ID:05c9a6ea-140a-49a2-bde5-f220b05aa252 Type:ContainerStarted Data:fbba8d600bd026ca0a23b1c9724088c18597af8b742d44a66208c8f6b0dbe0ad} Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:42.912471 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event=&{ID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Type:ContainerStarted Data:3676f1603de80edd23f34248eef289ffee599487f3f97ddd83f3362b1b7dc2f0} Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:42.913743 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-2pfzf" event=&{ID:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 Type:ContainerStarted Data:f218c2963bb329ec34cc051d3d32e91b0f36d73cbc9ea90fa4cdef808dbf1e48} Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:42.914876 2927 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-2pfzf" Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:42.916678 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" event=&{ID:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e Type:ContainerStarted Data:047cddfaab1d98f2f4be03b086b5acc263c7285299eed9b70a12b9c6485a6435} Jan 05 09:29:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:42.922574 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" event=&{ID:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 Type:ContainerStarted Data:b1800c468fe04ef1ee6fc5ecf08f9b47946f276d072facfdc0f9944871c84001} Jan 05 09:29:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:43.174132 2927 patch_prober.go:29] interesting pod/router-default-5cf6bbdff9-tnffs container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 05 09:29:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: [-]has-synced failed: reason withheld Jan 05 09:29:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: [+]process-running ok Jan 05 09:29:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: healthz check failed Jan 05 09:29:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:43.174194 2927 prober.go:114] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" podUID=c3b10ff4-a243-4a4e-b27d-f613e258666d containerName="router" probeResult=failure output="HTTP probe failed with statuscode: 500" Jan 05 09:29:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:43.198252946Z" level=info msg="Created container ff00bd198c611d428013e67929b1264e892e200f10b113006fa289c79f2c6c50: openshift-monitoring/prometheus-k8s-0/thanos-sidecar" id=5a44d88d-4bcf-4e20-ad44-59302e23377e name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:43.202373343Z" level=info msg="Starting container: ff00bd198c611d428013e67929b1264e892e200f10b113006fa289c79f2c6c50" id=ff427291-ce0c-486b-a682-f4603a82b8ee name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:43.271995107Z" level=info msg="Started container" PID=7655 containerID=ff00bd198c611d428013e67929b1264e892e200f10b113006fa289c79f2c6c50 description=openshift-monitoring/prometheus-k8s-0/thanos-sidecar id=ff427291-ce0c-486b-a682-f4603a82b8ee name=/runtime.v1.RuntimeService/StartContainer sandboxID=04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0 Jan 05 09:29:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:43.275064985Z" level=info msg="Created container f4fe2d6c3cc9d0c0aa0e03b9983f735dd906f6abbf9c60a8e821236675195f71: openshift-monitoring/alertmanager-main-0/prom-label-proxy" id=290cf590-b704-4f51-8012-cc48cb78b694 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:43.281030082Z" level=info msg="Starting container: f4fe2d6c3cc9d0c0aa0e03b9983f735dd906f6abbf9c60a8e821236675195f71" id=73388655-8b9c-49e7-8c03-c9d304914b61 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:43.304487232Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=e0781977-59ef-47c1-8051-0cffd935412a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:43.304707378Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=e0781977-59ef-47c1-8051-0cffd935412a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:43.305576491Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=c2047d7a-bddd-4b93-a1b5-78403b669584 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:43.305769493Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=c2047d7a-bddd-4b93-a1b5-78403b669584 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:43.306847537Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-0/prometheus-proxy" id=f0042379-48ce-42a6-9fda-769a77b6b89c name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:43.306986198Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:43.335196985Z" level=info msg="Started container" PID=7653 containerID=f4fe2d6c3cc9d0c0aa0e03b9983f735dd906f6abbf9c60a8e821236675195f71 description=openshift-monitoring/alertmanager-main-0/prom-label-proxy id=73388655-8b9c-49e7-8c03-c9d304914b61 name=/runtime.v1.RuntimeService/StartContainer sandboxID=de5e12e9744f461b9b6c0146346e5117bb31a1f6538299e5dccde6d8250c4268 Jan 05 09:29:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:43.716768443Z" level=info msg="Created container 7de1317b2bf9c95059be217ab261be612eec0edba15cda55d2880d2366374632: openshift-monitoring/prometheus-k8s-0/prometheus-proxy" id=f0042379-48ce-42a6-9fda-769a77b6b89c name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:43.717855667Z" level=info msg="Starting container: 7de1317b2bf9c95059be217ab261be612eec0edba15cda55d2880d2366374632" id=aa8e04ec-f280-4823-bedd-84666c690fb2 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:43.743550521Z" level=info msg="Started container" PID=7721 containerID=7de1317b2bf9c95059be217ab261be612eec0edba15cda55d2880d2366374632 description=openshift-monitoring/prometheus-k8s-0/prometheus-proxy id=aa8e04ec-f280-4823-bedd-84666c690fb2 name=/runtime.v1.RuntimeService/StartContainer sandboxID=04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0 Jan 05 09:29:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:43.788733205Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=89cc12ce-c087-4f61-b903-5d74ac772094 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:43.789009494Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=89cc12ce-c087-4f61-b903-5d74ac772094 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:43.790147862Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=daebe879-4fec-4ce5-a118-46291108223f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:43.790362131Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=daebe879-4fec-4ce5-a118-46291108223f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:43.791346880Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-0/kube-rbac-proxy" id=f47627df-321c-4699-8c0e-39f21480b06a name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:43.791453310Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:43.955623 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event=&{ID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Type:ContainerStarted Data:7de1317b2bf9c95059be217ab261be612eec0edba15cda55d2880d2366374632} Jan 05 09:29:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:43.955659 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event=&{ID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Type:ContainerStarted Data:ff00bd198c611d428013e67929b1264e892e200f10b113006fa289c79f2c6c50} Jan 05 09:29:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:43.968197 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event=&{ID:05c9a6ea-140a-49a2-bde5-f220b05aa252 Type:ContainerStarted Data:f4fe2d6c3cc9d0c0aa0e03b9983f735dd906f6abbf9c60a8e821236675195f71} Jan 05 09:29:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:44.168005 2927 patch_prober.go:29] interesting pod/router-default-5cf6bbdff9-tnffs container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 05 09:29:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: [-]has-synced failed: reason withheld Jan 05 09:29:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: [+]process-running ok Jan 05 09:29:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: healthz check failed Jan 05 09:29:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:44.168062 2927 prober.go:114] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" podUID=c3b10ff4-a243-4a4e-b27d-f613e258666d containerName="router" probeResult=failure output="HTTP probe failed with statuscode: 500" Jan 05 09:29:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:44.179025346Z" level=info msg="Created container 5e35b3bbc117459ece96a932475aecb63a3482c57612aebbe51e956cad2125e9: openshift-monitoring/prometheus-k8s-0/kube-rbac-proxy" id=f47627df-321c-4699-8c0e-39f21480b06a name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:44.179473927Z" level=info msg="Starting container: 5e35b3bbc117459ece96a932475aecb63a3482c57612aebbe51e956cad2125e9" id=1d4daccb-b868-488c-a137-d17611513455 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:44.214556083Z" level=info msg="Started container" PID=7768 containerID=5e35b3bbc117459ece96a932475aecb63a3482c57612aebbe51e956cad2125e9 description=openshift-monitoring/prometheus-k8s-0/kube-rbac-proxy id=1d4daccb-b868-488c-a137-d17611513455 name=/runtime.v1.RuntimeService/StartContainer sandboxID=04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0 Jan 05 09:29:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:44.246655183Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=a1dd27b2-a0e7-488f-a5bb-ac7917a34944 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:44.246881748Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=a1dd27b2-a0e7-488f-a5bb-ac7917a34944 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:44.248044291Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=57eeccd9-81e4-4946-b33c-a8a7adcb0848 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:44.248222745Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=57eeccd9-81e4-4946-b33c-a8a7adcb0848 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:44.249206584Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-0/kube-rbac-proxy-thanos" id=8ac516e8-44a6-4347-a66a-7f0cbd3fcf5e name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:44.249314312Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:44.623085693Z" level=info msg="Created container 970a12721ad9e883397cc47eca2c43e0043682e646f53f4100c9f77099dd9974: openshift-monitoring/prometheus-k8s-0/kube-rbac-proxy-thanos" id=8ac516e8-44a6-4347-a66a-7f0cbd3fcf5e name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:44.631569198Z" level=info msg="Starting container: 970a12721ad9e883397cc47eca2c43e0043682e646f53f4100c9f77099dd9974" id=e8a94210-6f7d-4c01-b204-9d8e34595e68 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:44.668110539Z" level=info msg="Started container" PID=7805 containerID=970a12721ad9e883397cc47eca2c43e0043682e646f53f4100c9f77099dd9974 description=openshift-monitoring/prometheus-k8s-0/kube-rbac-proxy-thanos id=e8a94210-6f7d-4c01-b204-9d8e34595e68 name=/runtime.v1.RuntimeService/StartContainer sandboxID=04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0 Jan 05 09:29:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:44.979360 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event=&{ID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Type:ContainerStarted Data:970a12721ad9e883397cc47eca2c43e0043682e646f53f4100c9f77099dd9974} Jan 05 09:29:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:44.979395 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event=&{ID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Type:ContainerStarted Data:5e35b3bbc117459ece96a932475aecb63a3482c57612aebbe51e956cad2125e9} Jan 05 09:29:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:45.156456 2927 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:29:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:45.169399 2927 patch_prober.go:29] interesting pod/router-default-5cf6bbdff9-tnffs container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 05 09:29:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: [-]has-synced failed: reason withheld Jan 05 09:29:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: [+]process-running ok Jan 05 09:29:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: healthz check failed Jan 05 09:29:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:45.169453 2927 prober.go:114] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" podUID=c3b10ff4-a243-4a4e-b27d-f613e258666d containerName="router" probeResult=failure output="HTTP probe failed with statuscode: 500" Jan 05 09:29:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:45.225989 2927 kubelet.go:2229] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:29:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:45.253590 2927 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:29:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:46.156301 2927 patch_prober.go:29] interesting pod/router-default-5cf6bbdff9-tnffs container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 05 09:29:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: [-]has-synced failed: reason withheld Jan 05 09:29:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: [+]process-running ok Jan 05 09:29:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: healthz check failed Jan 05 09:29:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:46.156355 2927 prober.go:114] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" podUID=c3b10ff4-a243-4a4e-b27d-f613e258666d containerName="router" probeResult=failure output="HTTP probe failed with statuscode: 500" Jan 05 09:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:47.155480 2927 patch_prober.go:29] interesting pod/router-default-5cf6bbdff9-tnffs container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 05 09:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: [-]has-synced failed: reason withheld Jan 05 09:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: [+]process-running ok Jan 05 09:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: healthz check failed Jan 05 09:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:47.155545 2927 prober.go:114] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" podUID=c3b10ff4-a243-4a4e-b27d-f613e258666d containerName="router" probeResult=failure output="HTTP probe failed with statuscode: 500" Jan 05 09:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:47.308123455Z" level=info msg="Got pod network &{Name:thanos-querier-6f4f5fb7cf-rdm9l Namespace:openshift-monitoring ID:f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3 UID:7c63c9be-d1ad-44b6-8b33-09af3ee314af NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:47.308305437Z" level=info msg="Deleting pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:47.308408723Z" level=error msg="Failed to cleanup (probably retrying): failed to destroy network for pod sandbox k8s_thanos-querier-6f4f5fb7cf-rdm9l_openshift-monitoring_7c63c9be-d1ad-44b6-8b33-09af3ee314af_0(f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3): error removing pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): netplugin failed with no error message: context deadline exceeded" Jan 05 09:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:47.308155916Z" level=error msg="Error stopping network on cleanup: failed to destroy network for pod sandbox k8s_thanos-querier-6f4f5fb7cf-rdm9l_openshift-monitoring_7c63c9be-d1ad-44b6-8b33-09af3ee314af_0(60f4edc9298bb10efa1323e01ae8e14c10e33a612f4d624c4f4ce9aabba94896): error removing pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/7c63c9be-d1ad-44b6-8b33-09af3ee314af]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" id=510e0b76-941b-4df9-96cf-95b4608a89ed name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:47.308470435Z" level=info msg="runSandbox: cleaning up namespaces after failing to run sandbox 60f4edc9298bb10efa1323e01ae8e14c10e33a612f4d624c4f4ce9aabba94896" id=510e0b76-941b-4df9-96cf-95b4608a89ed name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:47.338969302Z" level=info msg="runSandbox: deleting pod ID 60f4edc9298bb10efa1323e01ae8e14c10e33a612f4d624c4f4ce9aabba94896 from idIndex" id=510e0b76-941b-4df9-96cf-95b4608a89ed name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:47.339011545Z" level=info msg="runSandbox: removing pod sandbox 60f4edc9298bb10efa1323e01ae8e14c10e33a612f4d624c4f4ce9aabba94896" id=510e0b76-941b-4df9-96cf-95b4608a89ed name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:47.339037131Z" level=info msg="runSandbox: deleting container ID from idIndex for sandbox 60f4edc9298bb10efa1323e01ae8e14c10e33a612f4d624c4f4ce9aabba94896" id=510e0b76-941b-4df9-96cf-95b4608a89ed name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:47.339057799Z" level=info msg="runSandbox: unmounting shmPath for sandbox 60f4edc9298bb10efa1323e01ae8e14c10e33a612f4d624c4f4ce9aabba94896" id=510e0b76-941b-4df9-96cf-95b4608a89ed name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:47.342956309Z" level=info msg="runSandbox: removing pod sandbox from storage: 60f4edc9298bb10efa1323e01ae8e14c10e33a612f4d624c4f4ce9aabba94896" id=510e0b76-941b-4df9-96cf-95b4608a89ed name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:47.353079123Z" level=info msg="runSandbox: releasing container name: k8s_POD_thanos-querier-6f4f5fb7cf-rdm9l_openshift-monitoring_7c63c9be-d1ad-44b6-8b33-09af3ee314af_0" id=510e0b76-941b-4df9-96cf-95b4608a89ed name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:47.353105460Z" level=info msg="runSandbox: releasing pod sandbox name: k8s_thanos-querier-6f4f5fb7cf-rdm9l_openshift-monitoring_7c63c9be-d1ad-44b6-8b33-09af3ee314af_0" id=510e0b76-941b-4df9-96cf-95b4608a89ed name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:47.353309 2927 remote_runtime.go:222] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_thanos-querier-6f4f5fb7cf-rdm9l_openshift-monitoring_7c63c9be-d1ad-44b6-8b33-09af3ee314af_0(60f4edc9298bb10efa1323e01ae8e14c10e33a612f4d624c4f4ce9aabba94896): error adding pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/7c63c9be-d1ad-44b6-8b33-09af3ee314af]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" Jan 05 09:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:47.353401 2927 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_thanos-querier-6f4f5fb7cf-rdm9l_openshift-monitoring_7c63c9be-d1ad-44b6-8b33-09af3ee314af_0(60f4edc9298bb10efa1323e01ae8e14c10e33a612f4d624c4f4ce9aabba94896): error adding pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/7c63c9be-d1ad-44b6-8b33-09af3ee314af]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:47.353443 2927 kuberuntime_manager.go:772] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_thanos-querier-6f4f5fb7cf-rdm9l_openshift-monitoring_7c63c9be-d1ad-44b6-8b33-09af3ee314af_0(60f4edc9298bb10efa1323e01ae8e14c10e33a612f4d624c4f4ce9aabba94896): error adding pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/7c63c9be-d1ad-44b6-8b33-09af3ee314af]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:29:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:29:47.353528 2927 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"thanos-querier-6f4f5fb7cf-rdm9l_openshift-monitoring(7c63c9be-d1ad-44b6-8b33-09af3ee314af)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"thanos-querier-6f4f5fb7cf-rdm9l_openshift-monitoring(7c63c9be-d1ad-44b6-8b33-09af3ee314af)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_thanos-querier-6f4f5fb7cf-rdm9l_openshift-monitoring_7c63c9be-d1ad-44b6-8b33-09af3ee314af_0(60f4edc9298bb10efa1323e01ae8e14c10e33a612f4d624c4f4ce9aabba94896): error adding pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus\\\" name=\\\"multus-cni-network\\\" failed (add): Multus: [openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/7c63c9be-d1ad-44b6-8b33-09af3ee314af]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/80-openshift-network.conf. pollimmediate error: timed out waiting for the condition\"" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" podUID=7c63c9be-d1ad-44b6-8b33-09af3ee314af Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:48.017565 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.017899935Z" level=info msg="Running pod sandbox: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/POD" id=4048475b-e623-4794-99c6-e3fde25041f6 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.017990738Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.040886384Z" level=info msg="Got pod network &{Name:thanos-querier-6f4f5fb7cf-rdm9l Namespace:openshift-monitoring ID:e0a041e795d1586c6307ac859c48d38830f232aa0328f519a4f2db991d42e38a UID:7c63c9be-d1ad-44b6-8b33-09af3ee314af NetNS:/var/run/netns/c196c4ae-f75c-4a2d-8f59-85c45ced434c Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.040967245Z" level=info msg="Adding pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:48.155679 2927 patch_prober.go:29] interesting pod/router-default-5cf6bbdff9-tnffs container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: [-]has-synced failed: reason withheld Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: [+]process-running ok Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: healthz check failed Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:48.155736 2927 prober.go:114] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" podUID=c3b10ff4-a243-4a4e-b27d-f613e258666d containerName="router" probeResult=failure output="HTTP probe failed with statuscode: 500" Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:48.412755 2927 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l] Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: I0105 09:29:48.196271 7902 ovs.go:90] Maximum command line arguments set to: 191102 Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: 2023-01-05T09:29:48Z [verbose] Add: openshift-monitoring:thanos-querier-6f4f5fb7cf-rdm9l:7c63c9be-d1ad-44b6-8b33-09af3ee314af:ovn-kubernetes(ovn-kubernetes):eth0 {"cniVersion":"0.4.0","interfaces":[{"name":"e0a041e795d1586","mac":"72:b9:c3:da:5a:58"},{"name":"eth0","mac":"0a:58:0a:83:00:10","sandbox":"/var/run/netns/c196c4ae-f75c-4a2d-8f59-85c45ced434c"}],"ips":[{"version":"4","interface":1,"address":"10.131.0.16/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: I0105 09:29:48.345615 7896 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"thanos-querier-6f4f5fb7cf-rdm9l", UID:"7c63c9be-d1ad-44b6-8b33-09af3ee314af", APIVersion:"v1", ResourceVersion:"90116", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.16/23] from ovn-kubernetes Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.424368895Z" level=info msg="Got pod network &{Name:thanos-querier-6f4f5fb7cf-rdm9l Namespace:openshift-monitoring ID:e0a041e795d1586c6307ac859c48d38830f232aa0328f519a4f2db991d42e38a UID:7c63c9be-d1ad-44b6-8b33-09af3ee314af NetNS:/var/run/netns/c196c4ae-f75c-4a2d-8f59-85c45ced434c Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.424553177Z" level=info msg="Checking pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l for CNI network multus-cni-network (type=multus)" Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:29:48.426943 2927 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c63c9be_d1ad_44b6_8b33_09af3ee314af.slice/crio-e0a041e795d1586c6307ac859c48d38830f232aa0328f519a4f2db991d42e38a.scope WatchSource:0}: Error finding container e0a041e795d1586c6307ac859c48d38830f232aa0328f519a4f2db991d42e38a: Status 404 returned error can't find the container with id e0a041e795d1586c6307ac859c48d38830f232aa0328f519a4f2db991d42e38a Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.434464470Z" level=info msg="Ran pod sandbox e0a041e795d1586c6307ac859c48d38830f232aa0328f519a4f2db991d42e38a with infra container: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/POD" id=4048475b-e623-4794-99c6-e3fde25041f6 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.437009858Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc" id=93896376-44c1-4059-a5e9-504f3aa4b251 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.437216642Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:08ff7a43d9185529520a29e114f0c28b070000b464a44b347240015f3b466488,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc],Size_:368278089,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=93896376-44c1-4059-a5e9-504f3aa4b251 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.438008700Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc" id=95648437-f5ce-4d2a-b1ce-a43cba2b88dd name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.438184560Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:08ff7a43d9185529520a29e114f0c28b070000b464a44b347240015f3b466488,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc],Size_:368278089,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=95648437-f5ce-4d2a-b1ce-a43cba2b88dd name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.439170648Z" level=info msg="Creating container: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/thanos-query" id=eb061464-9734-4904-9911-790f1bbe4166 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.439279749Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.584262940Z" level=info msg="Created container 3d221c9b7969a4393326f6db0244fa128281bf5eb98e6da040c6d3d1135a666d: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/thanos-query" id=eb061464-9734-4904-9911-790f1bbe4166 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.584639834Z" level=info msg="Starting container: 3d221c9b7969a4393326f6db0244fa128281bf5eb98e6da040c6d3d1135a666d" id=fa76fd7e-9fa7-437f-a6de-597dae4743e8 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.593407164Z" level=info msg="Started container" PID=7946 containerID=3d221c9b7969a4393326f6db0244fa128281bf5eb98e6da040c6d3d1135a666d description=openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/thanos-query id=fa76fd7e-9fa7-437f-a6de-597dae4743e8 name=/runtime.v1.RuntimeService/StartContainer sandboxID=e0a041e795d1586c6307ac859c48d38830f232aa0328f519a4f2db991d42e38a Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.606137441Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=a5edcc32-b73a-41e7-acad-9afeddc79b8b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.606383388Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=a5edcc32-b73a-41e7-acad-9afeddc79b8b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.607554050Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=ad783040-358d-44fa-9b39-36ca726cdcda name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.607784642Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=ad783040-358d-44fa-9b39-36ca726cdcda name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.609053252Z" level=info msg="Creating container: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/oauth-proxy" id=fef98949-5b29-40d3-b571-e0e5774ea276 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.609164221Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.771571374Z" level=info msg="Created container 00f696cbb09d5836744a030a84da636141f60f03f340f69b27a9677ead91bf17: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/oauth-proxy" id=fef98949-5b29-40d3-b571-e0e5774ea276 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.771993087Z" level=info msg="Starting container: 00f696cbb09d5836744a030a84da636141f60f03f340f69b27a9677ead91bf17" id=aa0f752f-49cc-4b1d-bab7-bf84f60e2dbf name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.780272092Z" level=info msg="Started container" PID=7987 containerID=00f696cbb09d5836744a030a84da636141f60f03f340f69b27a9677ead91bf17 description=openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/oauth-proxy id=aa0f752f-49cc-4b1d-bab7-bf84f60e2dbf name=/runtime.v1.RuntimeService/StartContainer sandboxID=e0a041e795d1586c6307ac859c48d38830f232aa0328f519a4f2db991d42e38a Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.799840188Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=1cdd2508-7bdd-4435-9a86-f2f130814bba name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.800112650Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=1cdd2508-7bdd-4435-9a86-f2f130814bba name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.801322320Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=1f839d6b-b69d-4456-937b-c6f39ce38695 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.801524759Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=1f839d6b-b69d-4456-937b-c6f39ce38695 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.803981091Z" level=info msg="Creating container: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy" id=c97b8df7-0e8f-4266-801f-b45e2333fe22 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.804098551Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.959724698Z" level=info msg="Created container 668fca10db53532d614671680bb989d57b924efcbca09e6ef2b7e4895ccd7abb: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy" id=c97b8df7-0e8f-4266-801f-b45e2333fe22 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.960175781Z" level=info msg="Starting container: 668fca10db53532d614671680bb989d57b924efcbca09e6ef2b7e4895ccd7abb" id=032c481d-de88-4f1d-a545-428a0e5f494b name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.968528808Z" level=info msg="Started container" PID=8032 containerID=668fca10db53532d614671680bb989d57b924efcbca09e6ef2b7e4895ccd7abb description=openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy id=032c481d-de88-4f1d-a545-428a0e5f494b name=/runtime.v1.RuntimeService/StartContainer sandboxID=e0a041e795d1586c6307ac859c48d38830f232aa0328f519a4f2db991d42e38a Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.985562447Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816" id=a831c543-a2a5-4d40-b7b0-db653f21d864 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.985792517Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:955ee2b41ecd88e3a9c0183a0b65d6edbf524cefa1318837dcbc82b0466c4d55,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816],Size_:309576938,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=a831c543-a2a5-4d40-b7b0-db653f21d864 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.986556684Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816" id=de191077-e4fd-47ae-ba59-9616a5569f89 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.986742452Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:955ee2b41ecd88e3a9c0183a0b65d6edbf524cefa1318837dcbc82b0466c4d55,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816],Size_:309576938,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=de191077-e4fd-47ae-ba59-9616a5569f89 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.987376753Z" level=info msg="Creating container: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/prom-label-proxy" id=c527be9c-94b7-4183-9c35-faa817e2ecdd name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:48.987481664Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:49.021203 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" event=&{ID:7c63c9be-d1ad-44b6-8b33-09af3ee314af Type:ContainerStarted Data:668fca10db53532d614671680bb989d57b924efcbca09e6ef2b7e4895ccd7abb} Jan 05 09:29:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:49.021242 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" event=&{ID:7c63c9be-d1ad-44b6-8b33-09af3ee314af Type:ContainerStarted Data:00f696cbb09d5836744a030a84da636141f60f03f340f69b27a9677ead91bf17} Jan 05 09:29:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:49.021259 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" event=&{ID:7c63c9be-d1ad-44b6-8b33-09af3ee314af Type:ContainerStarted Data:3d221c9b7969a4393326f6db0244fa128281bf5eb98e6da040c6d3d1135a666d} Jan 05 09:29:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:49.021275 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" event=&{ID:7c63c9be-d1ad-44b6-8b33-09af3ee314af Type:ContainerStarted Data:e0a041e795d1586c6307ac859c48d38830f232aa0328f519a4f2db991d42e38a} Jan 05 09:29:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:49.126374222Z" level=info msg="Created container ba948b0e959411da071a39a920e3e612512d962feea740e6d05b34f3da0f9919: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/prom-label-proxy" id=c527be9c-94b7-4183-9c35-faa817e2ecdd name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:49.126771081Z" level=info msg="Starting container: ba948b0e959411da071a39a920e3e612512d962feea740e6d05b34f3da0f9919" id=42d23a6c-f461-44a3-98a5-e7090da94fd8 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:49.135446107Z" level=info msg="Started container" PID=8069 containerID=ba948b0e959411da071a39a920e3e612512d962feea740e6d05b34f3da0f9919 description=openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/prom-label-proxy id=42d23a6c-f461-44a3-98a5-e7090da94fd8 name=/runtime.v1.RuntimeService/StartContainer sandboxID=e0a041e795d1586c6307ac859c48d38830f232aa0328f519a4f2db991d42e38a Jan 05 09:29:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:49.155497 2927 kubelet.go:2229] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:29:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:49.158500 2927 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:29:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:49.166067308Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=7b5cbe9d-d035-474b-8250-b03161369f7e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:49.166303292Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=7b5cbe9d-d035-474b-8250-b03161369f7e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:49.167223953Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=05d82593-6682-4074-879a-d42e67a47fdc name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:49.167411185Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=05d82593-6682-4074-879a-d42e67a47fdc name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:49.168532780Z" level=info msg="Creating container: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy-rules" id=d7c9c702-1171-4d89-b141-ef2cd0f16f9a name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:49.168652635Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:49.349883458Z" level=info msg="Created container 205169b62d7080fa5dcfe7076f14dc28577ed4c769ffe43d35fb73df48f1cd2e: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy-rules" id=d7c9c702-1171-4d89-b141-ef2cd0f16f9a name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:49.350294721Z" level=info msg="Starting container: 205169b62d7080fa5dcfe7076f14dc28577ed4c769ffe43d35fb73df48f1cd2e" id=b7aebc95-d22c-4ddb-bdcc-84d58432d493 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:49.359019099Z" level=info msg="Started container" PID=8108 containerID=205169b62d7080fa5dcfe7076f14dc28577ed4c769ffe43d35fb73df48f1cd2e description=openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy-rules id=b7aebc95-d22c-4ddb-bdcc-84d58432d493 name=/runtime.v1.RuntimeService/StartContainer sandboxID=e0a041e795d1586c6307ac859c48d38830f232aa0328f519a4f2db991d42e38a Jan 05 09:29:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:49.375689145Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=78649d1a-ee9e-4d56-bffd-a55a691b77e9 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:49.375981171Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=78649d1a-ee9e-4d56-bffd-a55a691b77e9 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:49.376801936Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=e871cb06-43d8-4a26-ae77-4776108cb86c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:49.377044365Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=e871cb06-43d8-4a26-ae77-4776108cb86c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:29:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:49.377840482Z" level=info msg="Creating container: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy-metrics" id=fdbe567a-885c-4dfa-9053-c1b89f761f4f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:49.377978188Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:29:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:49.566169740Z" level=info msg="Created container f4d6579886d4371779b80ec613338f22c6a967a0e93854fba505389a0fc61b57: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy-metrics" id=fdbe567a-885c-4dfa-9053-c1b89f761f4f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:29:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:49.566543835Z" level=info msg="Starting container: f4d6579886d4371779b80ec613338f22c6a967a0e93854fba505389a0fc61b57" id=36dbe646-6393-4c6f-82d1-bc17e5b92f6f name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:29:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:49.599065432Z" level=info msg="Started container" PID=8146 containerID=f4d6579886d4371779b80ec613338f22c6a967a0e93854fba505389a0fc61b57 description=openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy-metrics id=36dbe646-6393-4c6f-82d1-bc17e5b92f6f name=/runtime.v1.RuntimeService/StartContainer sandboxID=e0a041e795d1586c6307ac859c48d38830f232aa0328f519a4f2db991d42e38a Jan 05 09:29:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:50.025795 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" event=&{ID:7c63c9be-d1ad-44b6-8b33-09af3ee314af Type:ContainerStarted Data:f4d6579886d4371779b80ec613338f22c6a967a0e93854fba505389a0fc61b57} Jan 05 09:29:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:50.025827 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" event=&{ID:7c63c9be-d1ad-44b6-8b33-09af3ee314af Type:ContainerStarted Data:205169b62d7080fa5dcfe7076f14dc28577ed4c769ffe43d35fb73df48f1cd2e} Jan 05 09:29:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:50.025843 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" event=&{ID:7c63c9be-d1ad-44b6-8b33-09af3ee314af Type:ContainerStarted Data:ba948b0e959411da071a39a920e3e612512d962feea740e6d05b34f3da0f9919} Jan 05 09:29:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:50.195480 2927 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:29:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:50.243123 2927 kubelet.go:2229] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:29:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:50.332485 2927 kubelet.go:2229] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:29:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:51.078206 2927 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:29:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:51.105998278Z" level=info msg="cleanup sandbox network" Jan 05 09:29:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:51.106328171Z" level=info msg="Got pod network &{Name:thanos-querier-6f4f5fb7cf-rdm9l Namespace:openshift-monitoring ID:f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3 UID:7c63c9be-d1ad-44b6-8b33-09af3ee314af NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:29:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:51.106476734Z" level=info msg="Deleting pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:29:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:53.379586 2927 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-2pfzf" Jan 05 09:29:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:29:55.206762 2927 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:29:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:29:57.638074683Z" level=warning msg="Found defunct process with PID 7927 (haproxy)" Jan 05 09:30:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:30:02.082011760Z" level=warning msg="Stopping container d4aa35750bb9cbb332c04ef2b3b110c8ccb79d5cb45a46daec15fffd82ab4b90 with stop signal timed out: timeout reached after 30 seconds waiting for container process to exit" id=59a32b47-1fb0-4108-8b67-1fbf4af6457c name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:30:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:30:02.348466 2927 remote_runtime.go:734] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4aa35750bb9cbb332c04ef2b3b110c8ccb79d5cb45a46daec15fffd82ab4b90 is running failed: container process not found" containerID="d4aa35750bb9cbb332c04ef2b3b110c8ccb79d5cb45a46daec15fffd82ab4b90" cmd=[/bin/bash -c test -f /ready/ready] Jan 05 09:30:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:30:02.348798 2927 remote_runtime.go:734] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4aa35750bb9cbb332c04ef2b3b110c8ccb79d5cb45a46daec15fffd82ab4b90 is running failed: container process not found" containerID="d4aa35750bb9cbb332c04ef2b3b110c8ccb79d5cb45a46daec15fffd82ab4b90" cmd=[/bin/bash -c test -f /ready/ready] Jan 05 09:30:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:30:02.349066 2927 remote_runtime.go:734] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4aa35750bb9cbb332c04ef2b3b110c8ccb79d5cb45a46daec15fffd82ab4b90 is running failed: container process not found" containerID="d4aa35750bb9cbb332c04ef2b3b110c8ccb79d5cb45a46daec15fffd82ab4b90" cmd=[/bin/bash -c test -f /ready/ready] Jan 05 09:30:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:30:02.349106 2927 prober.go:111] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d4aa35750bb9cbb332c04ef2b3b110c8ccb79d5cb45a46daec15fffd82ab4b90 is running failed: container process not found" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-ccgld" podUID=42d075ab-9d7c-4a5d-8c34-2d2b27434266 containerName="kube-multus-additional-cni-plugins" Jan 05 09:30:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:30:02.757039871Z" level=info msg="Stopped container d4aa35750bb9cbb332c04ef2b3b110c8ccb79d5cb45a46daec15fffd82ab4b90: openshift-multus/cni-sysctl-allowlist-ds-ccgld/kube-multus-additional-cni-plugins" id=59a32b47-1fb0-4108-8b67-1fbf4af6457c name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:30:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:30:02.758031672Z" level=info msg="Stopping pod sandbox: b0357b2427dbf428186e352b2d924d648a9f435c4ee4f03812d158cd89c3b229" id=d4355768-56be-4217-abf2-ca37298ed2c1 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:30:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:30:02.758320796Z" level=info msg="Got pod network &{Name:cni-sysctl-allowlist-ds-ccgld Namespace:openshift-multus ID:b0357b2427dbf428186e352b2d924d648a9f435c4ee4f03812d158cd89c3b229 UID:42d075ab-9d7c-4a5d-8c34-2d2b27434266 NetNS:/var/run/netns/a211c907-eb76-4818-b869-f0533d0c01e0 Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:30:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:30:02.758469675Z" level=info msg="Deleting pod openshift-multus_cni-sysctl-allowlist-ds-ccgld from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:30:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:30:03.064715 2927 generic.go:296] "Generic (PLEG): container finished" podID=42d075ab-9d7c-4a5d-8c34-2d2b27434266 containerID="d4aa35750bb9cbb332c04ef2b3b110c8ccb79d5cb45a46daec15fffd82ab4b90" exitCode=137 Jan 05 09:30:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:30:03.064762 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-ccgld" event=&{ID:42d075ab-9d7c-4a5d-8c34-2d2b27434266 Type:ContainerDied Data:d4aa35750bb9cbb332c04ef2b3b110c8ccb79d5cb45a46daec15fffd82ab4b90} Jan 05 09:30:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: 2023-01-05T09:30:02Z [verbose] Del: openshift-multus:cni-sysctl-allowlist-ds-ccgld:42d075ab-9d7c-4a5d-8c34-2d2b27434266:ovn-kubernetes:eth0 {"cniVersion":"0.4.0","dns":{},"ipam":{},"logFile":"/var/log/ovn-kubernetes/ovn-k8s-cni-overlay.log","logLevel":"4","logfile-maxage":5,"logfile-maxbackups":5,"logfile-maxsize":100,"name":"ovn-kubernetes","type":"ovn-k8s-cni-overlay"} Jan 05 09:30:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: I0105 09:30:02.968660 8520 ovs.go:90] Maximum command line arguments set to: 191102 Jan 05 09:30:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:30:03.126086269Z" level=info msg="Stopped pod sandbox: b0357b2427dbf428186e352b2d924d648a9f435c4ee4f03812d158cd89c3b229" id=d4355768-56be-4217-abf2-ca37298ed2c1 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:30:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:30:03.143385 2927 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zd8c\" (UniqueName: \"kubernetes.io/projected/42d075ab-9d7c-4a5d-8c34-2d2b27434266-kube-api-access-7zd8c\") pod \"42d075ab-9d7c-4a5d-8c34-2d2b27434266\" (UID: \"42d075ab-9d7c-4a5d-8c34-2d2b27434266\") " Jan 05 09:30:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:30:03.143432 2927 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/42d075ab-9d7c-4a5d-8c34-2d2b27434266-ready\") pod \"42d075ab-9d7c-4a5d-8c34-2d2b27434266\" (UID: \"42d075ab-9d7c-4a5d-8c34-2d2b27434266\") " Jan 05 09:30:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:30:03.143468 2927 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/42d075ab-9d7c-4a5d-8c34-2d2b27434266-tuning-conf-dir\") pod \"42d075ab-9d7c-4a5d-8c34-2d2b27434266\" (UID: \"42d075ab-9d7c-4a5d-8c34-2d2b27434266\") " Jan 05 09:30:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:30:03.143528 2927 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/42d075ab-9d7c-4a5d-8c34-2d2b27434266-cni-sysctl-allowlist\") pod \"42d075ab-9d7c-4a5d-8c34-2d2b27434266\" (UID: \"42d075ab-9d7c-4a5d-8c34-2d2b27434266\") " Jan 05 09:30:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:30:03.143790 2927 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/42d075ab-9d7c-4a5d-8c34-2d2b27434266/volumes/kubernetes.io~empty-dir/ready: clearQuota called, but quotas disabled Jan 05 09:30:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:30:03.143792 2927 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/42d075ab-9d7c-4a5d-8c34-2d2b27434266/volumes/kubernetes.io~configmap/cni-sysctl-allowlist: clearQuota called, but quotas disabled Jan 05 09:30:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:30:03.143881 2927 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42d075ab-9d7c-4a5d-8c34-2d2b27434266-ready" (OuterVolumeSpecName: "ready") pod "42d075ab-9d7c-4a5d-8c34-2d2b27434266" (UID: "42d075ab-9d7c-4a5d-8c34-2d2b27434266"). InnerVolumeSpecName "ready". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 05 09:30:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:30:03.144027 2927 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42d075ab-9d7c-4a5d-8c34-2d2b27434266-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "42d075ab-9d7c-4a5d-8c34-2d2b27434266" (UID: "42d075ab-9d7c-4a5d-8c34-2d2b27434266"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:30:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:30:03.144061 2927 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/42d075ab-9d7c-4a5d-8c34-2d2b27434266-tuning-conf-dir" (OuterVolumeSpecName: "tuning-conf-dir") pod "42d075ab-9d7c-4a5d-8c34-2d2b27434266" (UID: "42d075ab-9d7c-4a5d-8c34-2d2b27434266"). InnerVolumeSpecName "tuning-conf-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:30:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:30:03.150237 2927 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42d075ab-9d7c-4a5d-8c34-2d2b27434266-kube-api-access-7zd8c" (OuterVolumeSpecName: "kube-api-access-7zd8c") pod "42d075ab-9d7c-4a5d-8c34-2d2b27434266" (UID: "42d075ab-9d7c-4a5d-8c34-2d2b27434266"). InnerVolumeSpecName "kube-api-access-7zd8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:30:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:30:03.244783 2927 reconciler.go:399] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/42d075ab-9d7c-4a5d-8c34-2d2b27434266-cni-sysctl-allowlist\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:30:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:30:03.244815 2927 reconciler.go:399] "Volume detached for volume \"kube-api-access-7zd8c\" (UniqueName: \"kubernetes.io/projected/42d075ab-9d7c-4a5d-8c34-2d2b27434266-kube-api-access-7zd8c\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:30:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:30:03.244831 2927 reconciler.go:399] "Volume detached for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/42d075ab-9d7c-4a5d-8c34-2d2b27434266-ready\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:30:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:30:03.244848 2927 reconciler.go:399] "Volume detached for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/42d075ab-9d7c-4a5d-8c34-2d2b27434266-tuning-conf-dir\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:30:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:30:04.068650 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-ccgld" event=&{ID:42d075ab-9d7c-4a5d-8c34-2d2b27434266 Type:ContainerDied Data:b0357b2427dbf428186e352b2d924d648a9f435c4ee4f03812d158cd89c3b229} Jan 05 09:30:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:30:04.068700 2927 scope.go:115] "RemoveContainer" containerID="d4aa35750bb9cbb332c04ef2b3b110c8ccb79d5cb45a46daec15fffd82ab4b90" Jan 05 09:30:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:30:04.074455719Z" level=info msg="Removing container: d4aa35750bb9cbb332c04ef2b3b110c8ccb79d5cb45a46daec15fffd82ab4b90" id=4d14d684-daee-4a27-a361-df5ee91872c5 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:30:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:30:04.100281679Z" level=info msg="Removed container d4aa35750bb9cbb332c04ef2b3b110c8ccb79d5cb45a46daec15fffd82ab4b90: openshift-multus/cni-sysctl-allowlist-ds-ccgld/kube-multus-additional-cni-plugins" id=4d14d684-daee-4a27-a361-df5ee91872c5 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:30:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:30:04.201462 2927 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-multus/cni-sysctl-allowlist-ds-ccgld] Jan 05 09:30:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:30:04.228481 2927 kubelet.go:2129] "SyncLoop REMOVE" source="api" pods=[openshift-multus/cni-sysctl-allowlist-ds-ccgld] Jan 05 09:30:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:30:04.938673634Z" level=info msg="Stopping pod sandbox: b0357b2427dbf428186e352b2d924d648a9f435c4ee4f03812d158cd89c3b229" id=4d1096f6-12fa-4961-874c-43eb4f60b86a name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:30:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:30:04.938718587Z" level=info msg="Stopped pod sandbox (already stopped): b0357b2427dbf428186e352b2d924d648a9f435c4ee4f03812d158cd89c3b229" id=4d1096f6-12fa-4961-874c-43eb4f60b86a name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:30:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:30:04.940778 2927 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=42d075ab-9d7c-4a5d-8c34-2d2b27434266 path="/var/lib/kubelet/pods/42d075ab-9d7c-4a5d-8c34-2d2b27434266/volumes" Jan 05 09:30:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:30:05.287215 2927 kubelet.go:2229] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:30:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:30:06.498398633Z" level=warning msg="Found defunct process with PID 8396 (haproxy)" Jan 05 09:30:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:30:10.175612 2927 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:30:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:30:15.467124 2927 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 09:30:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:30:27.634900312Z" level=warning msg="Found defunct process with PID 8689 (haproxy)" Jan 05 09:30:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:30:27.635025292Z" level=warning msg="Found defunct process with PID 8944 (haproxy)" Jan 05 09:30:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:30:28.707632220Z" level=info msg="Stopping pod sandbox: b0357b2427dbf428186e352b2d924d648a9f435c4ee4f03812d158cd89c3b229" id=c40d3254-7dd3-482b-8240-cfaae96e9937 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:30:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:30:28.707671834Z" level=info msg="Stopped pod sandbox (already stopped): b0357b2427dbf428186e352b2d924d648a9f435c4ee4f03812d158cd89c3b229" id=c40d3254-7dd3-482b-8240-cfaae96e9937 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:30:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:30:28.708027999Z" level=info msg="Removing pod sandbox: b0357b2427dbf428186e352b2d924d648a9f435c4ee4f03812d158cd89c3b229" id=51763d5f-acd4-48f9-8472-b2613dee240b name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:30:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:30:28.719346813Z" level=info msg="Removed pod sandbox: b0357b2427dbf428186e352b2d924d648a9f435c4ee4f03812d158cd89c3b229" id=51763d5f-acd4-48f9-8472-b2613dee240b name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:30:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:30:36.121037414Z" level=error msg="Failed to cleanup (probably retrying): failed to destroy network for pod sandbox k8s_thanos-querier-6f4f5fb7cf-rdm9l_openshift-monitoring_7c63c9be-d1ad-44b6-8b33-09af3ee314af_0(f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3): error removing pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/7c63c9be-d1ad-44b6-8b33-09af3ee314af]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" Jan 05 09:30:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:30:36.493343370Z" level=warning msg="Found defunct process with PID 9040 (haproxy)" Jan 05 09:30:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:30:41.817021758Z" level=info msg="cleanup sandbox network" Jan 05 09:30:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:30:41.817332634Z" level=info msg="Got pod network &{Name:thanos-querier-6f4f5fb7cf-rdm9l Namespace:openshift-monitoring ID:f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3 UID:7c63c9be-d1ad-44b6-8b33-09af3ee314af NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:30:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:30:41.817484482Z" level=info msg="Deleting pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:31:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:31:26.831512869Z" level=error msg="Failed to cleanup (probably retrying): failed to destroy network for pod sandbox k8s_thanos-querier-6f4f5fb7cf-rdm9l_openshift-monitoring_7c63c9be-d1ad-44b6-8b33-09af3ee314af_0(f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3): error removing pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/7c63c9be-d1ad-44b6-8b33-09af3ee314af]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" Jan 05 09:31:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:31:35.375456792Z" level=info msg="cleanup sandbox network" Jan 05 09:31:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:31:35.375774673Z" level=info msg="Got pod network &{Name:thanos-querier-6f4f5fb7cf-rdm9l Namespace:openshift-monitoring ID:f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3 UID:7c63c9be-d1ad-44b6-8b33-09af3ee314af NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:31:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:31:35.375951129Z" level=info msg="Deleting pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:31:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:31:36.493512143Z" level=warning msg="Found defunct process with PID 8546 (haproxy)" Jan 05 09:32:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:32:20.389013573Z" level=error msg="Failed to cleanup (probably retrying): failed to destroy network for pod sandbox k8s_thanos-querier-6f4f5fb7cf-rdm9l_openshift-monitoring_7c63c9be-d1ad-44b6-8b33-09af3ee314af_0(f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3): error removing pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/7c63c9be-d1ad-44b6-8b33-09af3ee314af]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" Jan 05 09:32:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:32:33.203749743Z" level=info msg="cleanup sandbox network" Jan 05 09:32:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:32:33.204086628Z" level=info msg="Got pod network &{Name:thanos-querier-6f4f5fb7cf-rdm9l Namespace:openshift-monitoring ID:f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3 UID:7c63c9be-d1ad-44b6-8b33-09af3ee314af NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:32:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:32:33.204249091Z" level=info msg="Deleting pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:33:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:33:18.215393694Z" level=error msg="Failed to cleanup (probably retrying): failed to destroy network for pod sandbox k8s_thanos-querier-6f4f5fb7cf-rdm9l_openshift-monitoring_7c63c9be-d1ad-44b6-8b33-09af3ee314af_0(f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3): error removing pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/7c63c9be-d1ad-44b6-8b33-09af3ee314af]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" Jan 05 09:33:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:33:37.437218101Z" level=info msg="cleanup sandbox network" Jan 05 09:33:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:33:37.437574363Z" level=info msg="Got pod network &{Name:thanos-querier-6f4f5fb7cf-rdm9l Namespace:openshift-monitoring ID:f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3 UID:7c63c9be-d1ad-44b6-8b33-09af3ee314af NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:33:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:33:37.437726833Z" level=info msg="Deleting pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:34:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:34:22.448251720Z" level=error msg="Failed to cleanup (probably retrying): failed to destroy network for pod sandbox k8s_thanos-querier-6f4f5fb7cf-rdm9l_openshift-monitoring_7c63c9be-d1ad-44b6-8b33-09af3ee314af_0(f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3): error removing pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/7c63c9be-d1ad-44b6-8b33-09af3ee314af]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" Jan 05 09:34:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:34:28.800018425Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=8740b7c4-b2ec-44db-a738-44bbe4da89fb name=/runtime.v1.ImageService/ImageStatus Jan 05 09:34:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:34:28.800292457Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=8740b7c4-b2ec-44db-a738-44bbe4da89fb name=/runtime.v1.ImageService/ImageStatus Jan 05 09:34:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:34:51.280944741Z" level=info msg="cleanup sandbox network" Jan 05 09:34:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:34:51.281300794Z" level=info msg="Got pod network &{Name:thanos-querier-6f4f5fb7cf-rdm9l Namespace:openshift-monitoring ID:f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3 UID:7c63c9be-d1ad-44b6-8b33-09af3ee314af NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:34:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:34:51.281451523Z" level=info msg="Deleting pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:34:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:56.026550 2927 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-ovn-kubernetes/ovnkube-node-rp9bd] Jan 05 09:34:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:56.026863 2927 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" podUID=aac1c4d4-278d-4a53-b2f7-a75ba43c2d18 containerName="ovn-controller" containerID="cri-o://2a002f48611de0f066f753f27aa97988bb851c0533c009a991efe9a325cfa713" gracePeriod=30 Jan 05 09:34:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:56.027284 2927 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" podUID=aac1c4d4-278d-4a53-b2f7-a75ba43c2d18 containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://d0db0990bab100d4dde6c13a68393ef82c94744cd6a2a6d42cf1f7fc68e5a389" gracePeriod=30 Jan 05 09:34:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:56.027449 2927 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" podUID=aac1c4d4-278d-4a53-b2f7-a75ba43c2d18 containerName="kube-rbac-proxy" containerID="cri-o://a942da66e3e5e6ad4915085345ab0c7d57f83d8c68551251fe4b168376e26596" gracePeriod=30 Jan 05 09:34:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:56.027515 2927 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" podUID=aac1c4d4-278d-4a53-b2f7-a75ba43c2d18 containerName="ovn-acl-logging" containerID="cri-o://a810bb6ca3335cd24a04949ad82533875c7fd3127e211ae9ebec6cea2c3dc045" gracePeriod=30 Jan 05 09:34:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:34:56.030002582Z" level=info msg="Stopping container: a810bb6ca3335cd24a04949ad82533875c7fd3127e211ae9ebec6cea2c3dc045 (timeout: 30s)" id=1c2f87dd-71e9-4287-9e19-92e8beae96e0 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:34:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:34:56.031693278Z" level=info msg="Stopping container: 2a002f48611de0f066f753f27aa97988bb851c0533c009a991efe9a325cfa713 (timeout: 30s)" id=e27bc86c-1539-4e23-afb1-8168f811dd96 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:34:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:34:56.031858035Z" level=info msg="Stopping container: d0db0990bab100d4dde6c13a68393ef82c94744cd6a2a6d42cf1f7fc68e5a389 (timeout: 30s)" id=2ff0004e-045c-4189-8746-a12aa2520df0 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:34:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:34:56.031992008Z" level=info msg="Stopping container: a942da66e3e5e6ad4915085345ab0c7d57f83d8c68551251fe4b168376e26596 (timeout: 30s)" id=c6844f01-b3ef-4ac2-9960-77455370158e name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:34:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:56.121281 2927 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" podUID=aac1c4d4-278d-4a53-b2f7-a75ba43c2d18 containerName="ovnkube-node" containerID="cri-o://23a50821ca44e2b87851e25e4b341565d308519d780b1b4bb88614b86cab434b" gracePeriod=30 Jan 05 09:34:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:34:56.121636485Z" level=info msg="Stopping container: 23a50821ca44e2b87851e25e4b341565d308519d780b1b4bb88614b86cab434b (timeout: 30s)" id=bfb701b1-58d0-4f7f-8235-bc17ce6a9452 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:34:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:34:56.181378172Z" level=info msg="Stopped container a810bb6ca3335cd24a04949ad82533875c7fd3127e211ae9ebec6cea2c3dc045: openshift-ovn-kubernetes/ovnkube-node-rp9bd/ovn-acl-logging" id=1c2f87dd-71e9-4287-9e19-92e8beae96e0 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:34:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:34:56.216483639Z" level=info msg="Stopped container 2a002f48611de0f066f753f27aa97988bb851c0533c009a991efe9a325cfa713: openshift-ovn-kubernetes/ovnkube-node-rp9bd/ovn-controller" id=e27bc86c-1539-4e23-afb1-8168f811dd96 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:34:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:34:56.285644566Z" level=info msg="Stopped container 23a50821ca44e2b87851e25e4b341565d308519d780b1b4bb88614b86cab434b: openshift-ovn-kubernetes/ovnkube-node-rp9bd/ovnkube-node" id=bfb701b1-58d0-4f7f-8235-bc17ce6a9452 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:34:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:56.976326 2927 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rp9bd_aac1c4d4-278d-4a53-b2f7-a75ba43c2d18/ovn-acl-logging/0.log" Jan 05 09:34:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:56.976894 2927 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rp9bd_aac1c4d4-278d-4a53-b2f7-a75ba43c2d18/ovn-controller/0.log" Jan 05 09:34:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:56.976955 2927 generic.go:296] "Generic (PLEG): container finished" podID=aac1c4d4-278d-4a53-b2f7-a75ba43c2d18 containerID="23a50821ca44e2b87851e25e4b341565d308519d780b1b4bb88614b86cab434b" exitCode=0 Jan 05 09:34:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:56.976973 2927 generic.go:296] "Generic (PLEG): container finished" podID=aac1c4d4-278d-4a53-b2f7-a75ba43c2d18 containerID="a810bb6ca3335cd24a04949ad82533875c7fd3127e211ae9ebec6cea2c3dc045" exitCode=143 Jan 05 09:34:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:56.976988 2927 generic.go:296] "Generic (PLEG): container finished" podID=aac1c4d4-278d-4a53-b2f7-a75ba43c2d18 containerID="2a002f48611de0f066f753f27aa97988bb851c0533c009a991efe9a325cfa713" exitCode=143 Jan 05 09:34:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:56.977013 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" event=&{ID:aac1c4d4-278d-4a53-b2f7-a75ba43c2d18 Type:ContainerDied Data:23a50821ca44e2b87851e25e4b341565d308519d780b1b4bb88614b86cab434b} Jan 05 09:34:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:56.977034 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" event=&{ID:aac1c4d4-278d-4a53-b2f7-a75ba43c2d18 Type:ContainerDied Data:a810bb6ca3335cd24a04949ad82533875c7fd3127e211ae9ebec6cea2c3dc045} Jan 05 09:34:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:56.977050 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" event=&{ID:aac1c4d4-278d-4a53-b2f7-a75ba43c2d18 Type:ContainerDied Data:2a002f48611de0f066f753f27aa97988bb851c0533c009a991efe9a325cfa713} Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:34:57.238016647Z" level=info msg="Stopped container a942da66e3e5e6ad4915085345ab0c7d57f83d8c68551251fe4b168376e26596: openshift-ovn-kubernetes/ovnkube-node-rp9bd/kube-rbac-proxy" id=c6844f01-b3ef-4ac2-9960-77455370158e name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:34:57.259044948Z" level=info msg="Stopped container d0db0990bab100d4dde6c13a68393ef82c94744cd6a2a6d42cf1f7fc68e5a389: openshift-ovn-kubernetes/ovnkube-node-rp9bd/kube-rbac-proxy-ovn-metrics" id=2ff0004e-045c-4189-8746-a12aa2520df0 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:34:57.259436746Z" level=info msg="Stopping pod sandbox: fd9deb131d1f752a07a6704193bcd1cf1dab19ff2e6439be2d911f94c338da51" id=9b12ea03-6798-48e5-ba9d-65e6fac538a1 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:34:57.305082881Z" level=info msg="Stopped pod sandbox: fd9deb131d1f752a07a6704193bcd1cf1dab19ff2e6439be2d911f94c338da51" id=9b12ea03-6798-48e5-ba9d-65e6fac538a1 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.313251 2927 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rp9bd_aac1c4d4-278d-4a53-b2f7-a75ba43c2d18/ovn-acl-logging/0.log" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.313696 2927 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rp9bd_aac1c4d4-278d-4a53-b2f7-a75ba43c2d18/ovn-controller/0.log" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.391671 2927 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-node-log\") pod \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.391719 2927 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-var-lib-cni-networks-ovn-kubernetes\") pod \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.391762 2927 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovn-node-metrics-cert\") pod \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.391787 2927 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-run-netns\") pod \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.391810 2927 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-cni-netd\") pod \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.391833 2927 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-slash\") pod \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.391872 2927 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovnkube-config\") pod \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.391911 2927 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"ovn-cert\" (UniqueName: \"kubernetes.io/secret/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovn-cert\") pod \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.391944 2927 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-log-socket\") pod \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.391978 2927 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-env-overrides\") pod \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.392418 2927 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-run-ovn-kubernetes\") pod \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.392481 2927 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-etc-openvswitch\") pod \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.392528 2927 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-var-lib-openvswitch\") pod \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.392570 2927 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"ovn-ca\" (UniqueName: \"kubernetes.io/configmap/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovn-ca\") pod \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.392606 2927 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-run-openvswitch\") pod \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.392637 2927 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-cni-bin\") pod \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.392683 2927 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csks4\" (UniqueName: \"kubernetes.io/projected/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-kube-api-access-csks4\") pod \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.392869 2927 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-systemd-units\") pod \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.392926 2927 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-run-ovn\") pod \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\" (UID: \"aac1c4d4-278d-4a53-b2f7-a75ba43c2d18\") " Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.393922 2927 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" (UID: "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.393968 2927 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-slash" (OuterVolumeSpecName: "host-slash") pod "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" (UID: "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:34:57.394175 2927 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18/volumes/kubernetes.io~configmap/ovnkube-config: clearQuota called, but quotas disabled Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.394394 2927 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" (UID: "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.396123 2927 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-node-log" (OuterVolumeSpecName: "node-log") pod "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" (UID: "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.396169 2927 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" (UID: "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.396493 2927 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" (UID: "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.396520 2927 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" (UID: "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.396557 2927 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" (UID: "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.396581 2927 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-log-socket" (OuterVolumeSpecName: "log-socket") pod "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" (UID: "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:34:57.396733 2927 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18/volumes/kubernetes.io~configmap/env-overrides: clearQuota called, but quotas disabled Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.396860 2927 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" (UID: "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.397585 2927 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" (UID: "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.397615 2927 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" (UID: "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.397646 2927 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" (UID: "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:34:57.397786 2927 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18/volumes/kubernetes.io~configmap/ovn-ca: clearQuota called, but quotas disabled Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.397973 2927 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovn-ca" (OuterVolumeSpecName: "ovn-ca") pod "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" (UID: "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18"). InnerVolumeSpecName "ovn-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.398008 2927 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" (UID: "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.398264 2927 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" (UID: "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.416063 2927 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovn-cert" (OuterVolumeSpecName: "ovn-cert") pod "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" (UID: "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18"). InnerVolumeSpecName "ovn-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.431055 2927 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" (UID: "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.439028 2927 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-kube-api-access-csks4" (OuterVolumeSpecName: "kube-api-access-csks4") pod "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" (UID: "aac1c4d4-278d-4a53-b2f7-a75ba43c2d18"). InnerVolumeSpecName "kube-api-access-csks4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.493883 2927 reconciler.go:399] "Volume detached for volume \"ovn-cert\" (UniqueName: \"kubernetes.io/secret/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovn-cert\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.493921 2927 reconciler.go:399] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovnkube-config\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.493937 2927 reconciler.go:399] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-log-socket\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.493953 2927 reconciler.go:399] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-env-overrides\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.493978 2927 reconciler.go:399] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-run-ovn-kubernetes\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.493994 2927 reconciler.go:399] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-etc-openvswitch\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.494009 2927 reconciler.go:399] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-run-openvswitch\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.494025 2927 reconciler.go:399] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-var-lib-openvswitch\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.494040 2927 reconciler.go:399] "Volume detached for volume \"ovn-ca\" (UniqueName: \"kubernetes.io/configmap/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovn-ca\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.494057 2927 reconciler.go:399] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-cni-bin\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.494074 2927 reconciler.go:399] "Volume detached for volume \"kube-api-access-csks4\" (UniqueName: \"kubernetes.io/projected/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-kube-api-access-csks4\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.494092 2927 reconciler.go:399] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-systemd-units\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.494108 2927 reconciler.go:399] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-run-ovn\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.494123 2927 reconciler.go:399] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-node-log\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.494142 2927 reconciler.go:399] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-var-lib-cni-networks-ovn-kubernetes\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.494160 2927 reconciler.go:399] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-ovn-node-metrics-cert\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.494176 2927 reconciler.go:399] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-slash\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.494194 2927 reconciler.go:399] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-run-netns\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.494224 2927 reconciler.go:399] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18-host-cni-netd\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.981176 2927 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rp9bd_aac1c4d4-278d-4a53-b2f7-a75ba43c2d18/ovn-acl-logging/0.log" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.981766 2927 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rp9bd_aac1c4d4-278d-4a53-b2f7-a75ba43c2d18/ovn-controller/0.log" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.981817 2927 generic.go:296] "Generic (PLEG): container finished" podID=aac1c4d4-278d-4a53-b2f7-a75ba43c2d18 containerID="d0db0990bab100d4dde6c13a68393ef82c94744cd6a2a6d42cf1f7fc68e5a389" exitCode=0 Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.981834 2927 generic.go:296] "Generic (PLEG): container finished" podID=aac1c4d4-278d-4a53-b2f7-a75ba43c2d18 containerID="a942da66e3e5e6ad4915085345ab0c7d57f83d8c68551251fe4b168376e26596" exitCode=0 Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.981862 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" event=&{ID:aac1c4d4-278d-4a53-b2f7-a75ba43c2d18 Type:ContainerDied Data:d0db0990bab100d4dde6c13a68393ef82c94744cd6a2a6d42cf1f7fc68e5a389} Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.981885 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" event=&{ID:aac1c4d4-278d-4a53-b2f7-a75ba43c2d18 Type:ContainerDied Data:a942da66e3e5e6ad4915085345ab0c7d57f83d8c68551251fe4b168376e26596} Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.981919 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rp9bd" event=&{ID:aac1c4d4-278d-4a53-b2f7-a75ba43c2d18 Type:ContainerDied Data:fd9deb131d1f752a07a6704193bcd1cf1dab19ff2e6439be2d911f94c338da51} Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:57.981940 2927 scope.go:115] "RemoveContainer" containerID="23a50821ca44e2b87851e25e4b341565d308519d780b1b4bb88614b86cab434b" Jan 05 09:34:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:34:57.988591458Z" level=info msg="Removing container: 23a50821ca44e2b87851e25e4b341565d308519d780b1b4bb88614b86cab434b" id=224f6e61-3ac2-4284-82ad-6c89b659a654 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:58.016588 2927 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-ovn-kubernetes/ovnkube-node-rp9bd] Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:58.022740 2927 kubelet.go:2129] "SyncLoop REMOVE" source="api" pods=[openshift-ovn-kubernetes/ovnkube-node-rp9bd] Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:34:58.024005939Z" level=info msg="Removed container 23a50821ca44e2b87851e25e4b341565d308519d780b1b4bb88614b86cab434b: openshift-ovn-kubernetes/ovnkube-node-rp9bd/ovnkube-node" id=224f6e61-3ac2-4284-82ad-6c89b659a654 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:58.025318 2927 scope.go:115] "RemoveContainer" containerID="d0db0990bab100d4dde6c13a68393ef82c94744cd6a2a6d42cf1f7fc68e5a389" Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:34:58.026053630Z" level=info msg="Removing container: d0db0990bab100d4dde6c13a68393ef82c94744cd6a2a6d42cf1f7fc68e5a389" id=b9ef6dc0-c941-4f68-acf1-cd63a467e22b name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:34:58.042860206Z" level=info msg="Removed container d0db0990bab100d4dde6c13a68393ef82c94744cd6a2a6d42cf1f7fc68e5a389: openshift-ovn-kubernetes/ovnkube-node-rp9bd/kube-rbac-proxy-ovn-metrics" id=b9ef6dc0-c941-4f68-acf1-cd63a467e22b name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:58.043031 2927 scope.go:115] "RemoveContainer" containerID="a942da66e3e5e6ad4915085345ab0c7d57f83d8c68551251fe4b168376e26596" Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:34:58.043709014Z" level=info msg="Removing container: a942da66e3e5e6ad4915085345ab0c7d57f83d8c68551251fe4b168376e26596" id=5d6adb05-df0a-478c-97d2-0fe490833add name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:34:58.068417022Z" level=info msg="Removed container a942da66e3e5e6ad4915085345ab0c7d57f83d8c68551251fe4b168376e26596: openshift-ovn-kubernetes/ovnkube-node-rp9bd/kube-rbac-proxy" id=5d6adb05-df0a-478c-97d2-0fe490833add name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:58.068620 2927 scope.go:115] "RemoveContainer" containerID="a810bb6ca3335cd24a04949ad82533875c7fd3127e211ae9ebec6cea2c3dc045" Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:34:58.069357626Z" level=info msg="Removing container: a810bb6ca3335cd24a04949ad82533875c7fd3127e211ae9ebec6cea2c3dc045" id=1be652df-1ce3-4d34-999b-f141ddfc244f name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:34:58.086126539Z" level=info msg="Removed container a810bb6ca3335cd24a04949ad82533875c7fd3127e211ae9ebec6cea2c3dc045: openshift-ovn-kubernetes/ovnkube-node-rp9bd/ovn-acl-logging" id=1be652df-1ce3-4d34-999b-f141ddfc244f name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:58.086303 2927 scope.go:115] "RemoveContainer" containerID="2a002f48611de0f066f753f27aa97988bb851c0533c009a991efe9a325cfa713" Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:34:58.086973612Z" level=info msg="Removing container: 2a002f48611de0f066f753f27aa97988bb851c0533c009a991efe9a325cfa713" id=52358bb6-e7bf-4004-98d0-42a1ecb1b4e8 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:34:58.104763953Z" level=info msg="Removed container 2a002f48611de0f066f753f27aa97988bb851c0533c009a991efe9a325cfa713: openshift-ovn-kubernetes/ovnkube-node-rp9bd/ovn-controller" id=52358bb6-e7bf-4004-98d0-42a1ecb1b4e8 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:58.104925 2927 scope.go:115] "RemoveContainer" containerID="23a50821ca44e2b87851e25e4b341565d308519d780b1b4bb88614b86cab434b" Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:34:58.105184 2927 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23a50821ca44e2b87851e25e4b341565d308519d780b1b4bb88614b86cab434b\": container with ID starting with 23a50821ca44e2b87851e25e4b341565d308519d780b1b4bb88614b86cab434b not found: ID does not exist" containerID="23a50821ca44e2b87851e25e4b341565d308519d780b1b4bb88614b86cab434b" Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:58.105240 2927 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:23a50821ca44e2b87851e25e4b341565d308519d780b1b4bb88614b86cab434b} err="failed to get container status \"23a50821ca44e2b87851e25e4b341565d308519d780b1b4bb88614b86cab434b\": rpc error: code = NotFound desc = could not find container \"23a50821ca44e2b87851e25e4b341565d308519d780b1b4bb88614b86cab434b\": container with ID starting with 23a50821ca44e2b87851e25e4b341565d308519d780b1b4bb88614b86cab434b not found: ID does not exist" Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:58.105255 2927 scope.go:115] "RemoveContainer" containerID="d0db0990bab100d4dde6c13a68393ef82c94744cd6a2a6d42cf1f7fc68e5a389" Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:34:58.105541 2927 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0db0990bab100d4dde6c13a68393ef82c94744cd6a2a6d42cf1f7fc68e5a389\": container with ID starting with d0db0990bab100d4dde6c13a68393ef82c94744cd6a2a6d42cf1f7fc68e5a389 not found: ID does not exist" containerID="d0db0990bab100d4dde6c13a68393ef82c94744cd6a2a6d42cf1f7fc68e5a389" Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:58.105575 2927 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:d0db0990bab100d4dde6c13a68393ef82c94744cd6a2a6d42cf1f7fc68e5a389} err="failed to get container status \"d0db0990bab100d4dde6c13a68393ef82c94744cd6a2a6d42cf1f7fc68e5a389\": rpc error: code = NotFound desc = could not find container \"d0db0990bab100d4dde6c13a68393ef82c94744cd6a2a6d42cf1f7fc68e5a389\": container with ID starting with d0db0990bab100d4dde6c13a68393ef82c94744cd6a2a6d42cf1f7fc68e5a389 not found: ID does not exist" Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:58.105589 2927 scope.go:115] "RemoveContainer" containerID="a942da66e3e5e6ad4915085345ab0c7d57f83d8c68551251fe4b168376e26596" Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:34:58.105798 2927 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a942da66e3e5e6ad4915085345ab0c7d57f83d8c68551251fe4b168376e26596\": container with ID starting with a942da66e3e5e6ad4915085345ab0c7d57f83d8c68551251fe4b168376e26596 not found: ID does not exist" containerID="a942da66e3e5e6ad4915085345ab0c7d57f83d8c68551251fe4b168376e26596" Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:58.105830 2927 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:a942da66e3e5e6ad4915085345ab0c7d57f83d8c68551251fe4b168376e26596} err="failed to get container status \"a942da66e3e5e6ad4915085345ab0c7d57f83d8c68551251fe4b168376e26596\": rpc error: code = NotFound desc = could not find container \"a942da66e3e5e6ad4915085345ab0c7d57f83d8c68551251fe4b168376e26596\": container with ID starting with a942da66e3e5e6ad4915085345ab0c7d57f83d8c68551251fe4b168376e26596 not found: ID does not exist" Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:58.105845 2927 scope.go:115] "RemoveContainer" containerID="a810bb6ca3335cd24a04949ad82533875c7fd3127e211ae9ebec6cea2c3dc045" Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:34:58.106074 2927 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a810bb6ca3335cd24a04949ad82533875c7fd3127e211ae9ebec6cea2c3dc045\": container with ID starting with a810bb6ca3335cd24a04949ad82533875c7fd3127e211ae9ebec6cea2c3dc045 not found: ID does not exist" containerID="a810bb6ca3335cd24a04949ad82533875c7fd3127e211ae9ebec6cea2c3dc045" Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:58.106106 2927 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:a810bb6ca3335cd24a04949ad82533875c7fd3127e211ae9ebec6cea2c3dc045} err="failed to get container status \"a810bb6ca3335cd24a04949ad82533875c7fd3127e211ae9ebec6cea2c3dc045\": rpc error: code = NotFound desc = could not find container \"a810bb6ca3335cd24a04949ad82533875c7fd3127e211ae9ebec6cea2c3dc045\": container with ID starting with a810bb6ca3335cd24a04949ad82533875c7fd3127e211ae9ebec6cea2c3dc045 not found: ID does not exist" Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:58.106120 2927 scope.go:115] "RemoveContainer" containerID="2a002f48611de0f066f753f27aa97988bb851c0533c009a991efe9a325cfa713" Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:34:58.106344 2927 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a002f48611de0f066f753f27aa97988bb851c0533c009a991efe9a325cfa713\": container with ID starting with 2a002f48611de0f066f753f27aa97988bb851c0533c009a991efe9a325cfa713 not found: ID does not exist" containerID="2a002f48611de0f066f753f27aa97988bb851c0533c009a991efe9a325cfa713" Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:58.106376 2927 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:2a002f48611de0f066f753f27aa97988bb851c0533c009a991efe9a325cfa713} err="failed to get container status \"2a002f48611de0f066f753f27aa97988bb851c0533c009a991efe9a325cfa713\": rpc error: code = NotFound desc = could not find container \"2a002f48611de0f066f753f27aa97988bb851c0533c009a991efe9a325cfa713\": container with ID starting with 2a002f48611de0f066f753f27aa97988bb851c0533c009a991efe9a325cfa713 not found: ID does not exist" Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:58.106388 2927 scope.go:115] "RemoveContainer" containerID="23a50821ca44e2b87851e25e4b341565d308519d780b1b4bb88614b86cab434b" Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:58.106594 2927 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:23a50821ca44e2b87851e25e4b341565d308519d780b1b4bb88614b86cab434b} err="failed to get container status \"23a50821ca44e2b87851e25e4b341565d308519d780b1b4bb88614b86cab434b\": rpc error: code = NotFound desc = could not find container \"23a50821ca44e2b87851e25e4b341565d308519d780b1b4bb88614b86cab434b\": container with ID starting with 23a50821ca44e2b87851e25e4b341565d308519d780b1b4bb88614b86cab434b not found: ID does not exist" Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:58.106609 2927 scope.go:115] "RemoveContainer" containerID="d0db0990bab100d4dde6c13a68393ef82c94744cd6a2a6d42cf1f7fc68e5a389" Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:58.106807 2927 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:d0db0990bab100d4dde6c13a68393ef82c94744cd6a2a6d42cf1f7fc68e5a389} err="failed to get container status \"d0db0990bab100d4dde6c13a68393ef82c94744cd6a2a6d42cf1f7fc68e5a389\": rpc error: code = NotFound desc = could not find container \"d0db0990bab100d4dde6c13a68393ef82c94744cd6a2a6d42cf1f7fc68e5a389\": container with ID starting with d0db0990bab100d4dde6c13a68393ef82c94744cd6a2a6d42cf1f7fc68e5a389 not found: ID does not exist" Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:58.106822 2927 scope.go:115] "RemoveContainer" containerID="a942da66e3e5e6ad4915085345ab0c7d57f83d8c68551251fe4b168376e26596" Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:58.107049 2927 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:a942da66e3e5e6ad4915085345ab0c7d57f83d8c68551251fe4b168376e26596} err="failed to get container status \"a942da66e3e5e6ad4915085345ab0c7d57f83d8c68551251fe4b168376e26596\": rpc error: code = NotFound desc = could not find container \"a942da66e3e5e6ad4915085345ab0c7d57f83d8c68551251fe4b168376e26596\": container with ID starting with a942da66e3e5e6ad4915085345ab0c7d57f83d8c68551251fe4b168376e26596 not found: ID does not exist" Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:58.107063 2927 scope.go:115] "RemoveContainer" containerID="a810bb6ca3335cd24a04949ad82533875c7fd3127e211ae9ebec6cea2c3dc045" Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:58.107282 2927 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:a810bb6ca3335cd24a04949ad82533875c7fd3127e211ae9ebec6cea2c3dc045} err="failed to get container status \"a810bb6ca3335cd24a04949ad82533875c7fd3127e211ae9ebec6cea2c3dc045\": rpc error: code = NotFound desc = could not find container \"a810bb6ca3335cd24a04949ad82533875c7fd3127e211ae9ebec6cea2c3dc045\": container with ID starting with a810bb6ca3335cd24a04949ad82533875c7fd3127e211ae9ebec6cea2c3dc045 not found: ID does not exist" Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:58.107303 2927 scope.go:115] "RemoveContainer" containerID="2a002f48611de0f066f753f27aa97988bb851c0533c009a991efe9a325cfa713" Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:58.107501 2927 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:2a002f48611de0f066f753f27aa97988bb851c0533c009a991efe9a325cfa713} err="failed to get container status \"2a002f48611de0f066f753f27aa97988bb851c0533c009a991efe9a325cfa713\": rpc error: code = NotFound desc = could not find container \"2a002f48611de0f066f753f27aa97988bb851c0533c009a991efe9a325cfa713\": container with ID starting with 2a002f48611de0f066f753f27aa97988bb851c0533c009a991efe9a325cfa713 not found: ID does not exist" Jan 05 09:34:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:34:58.941602 2927 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=aac1c4d4-278d-4a53-b2f7-a75ba43c2d18 path="/var/lib/kubelet/pods/aac1c4d4-278d-4a53-b2f7-a75ba43c2d18/volumes" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.108174 2927 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-sdn/sdn-66nsp] Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.108238 2927 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:35:05.108315 2927 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="42d075ab-9d7c-4a5d-8c34-2d2b27434266" containerName="kube-multus-additional-cni-plugins" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.108328 2927 state_mem.go:107] "Deleted CPUSet assignment" podUID="42d075ab-9d7c-4a5d-8c34-2d2b27434266" containerName="kube-multus-additional-cni-plugins" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:35:05.108344 2927 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" containerName="ovn-acl-logging" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.108352 2927 state_mem.go:107] "Deleted CPUSet assignment" podUID="aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" containerName="ovn-acl-logging" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:35:05.108365 2927 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" containerName="kube-rbac-proxy-ovn-metrics" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.108373 2927 state_mem.go:107] "Deleted CPUSet assignment" podUID="aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" containerName="kube-rbac-proxy-ovn-metrics" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:35:05.108386 2927 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" containerName="ovn-controller" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.108396 2927 state_mem.go:107] "Deleted CPUSet assignment" podUID="aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" containerName="ovn-controller" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:35:05.108406 2927 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" containerName="kube-rbac-proxy" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.108414 2927 state_mem.go:107] "Deleted CPUSet assignment" podUID="aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" containerName="kube-rbac-proxy" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:35:05.108427 2927 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" containerName="ovnkube-node" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.108435 2927 state_mem.go:107] "Deleted CPUSet assignment" podUID="aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" containerName="ovnkube-node" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.108502 2927 memory_manager.go:345] "RemoveStaleState removing state" podUID="aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" containerName="ovnkube-node" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.108513 2927 memory_manager.go:345] "RemoveStaleState removing state" podUID="42d075ab-9d7c-4a5d-8c34-2d2b27434266" containerName="kube-multus-additional-cni-plugins" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.108530 2927 memory_manager.go:345] "RemoveStaleState removing state" podUID="aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" containerName="kube-rbac-proxy" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.108540 2927 memory_manager.go:345] "RemoveStaleState removing state" podUID="aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" containerName="ovn-acl-logging" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.108550 2927 memory_manager.go:345] "RemoveStaleState removing state" podUID="aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" containerName="kube-rbac-proxy-ovn-metrics" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.108562 2927 memory_manager.go:345] "RemoveStaleState removing state" podUID="aac1c4d4-278d-4a53-b2f7-a75ba43c2d18" containerName="ovn-controller" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.148352 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5b0f5e6-3d70-4315-b9c5-1403036f517b-config\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.148399 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-mtu-migration\" (UniqueName: \"kubernetes.io/configmap/d5b0f5e6-3d70-4315-b9c5-1403036f517b-config-mtu-migration\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.148431 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-etc-sysconfig\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.148463 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-slash\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.148498 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nd4t\" (UniqueName: \"kubernetes.io/projected/d5b0f5e6-3d70-4315-b9c5-1403036f517b-kube-api-access-8nd4t\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.148528 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-run-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-kubernetes\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.148558 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-modules\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-modules\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.148594 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-run-ovs\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-ovs\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.148632 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d5b0f5e6-3d70-4315-b9c5-1403036f517b-env-overrides\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.148663 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-cni-bin\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.148696 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-lib-cni-networks-openshift-sdn\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.148730 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-conf\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-cni-conf\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.148759 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-dbus\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.148790 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-run-netns\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.148827 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-run-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-openshift-sdn\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.148860 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sdn-metrics-certs\" (UniqueName: \"kubernetes.io/secret/d5b0f5e6-3d70-4315-b9c5-1403036f517b-sdn-metrics-certs\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.148894 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-run-crio\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-crio\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.252329 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-run-netns\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.252383 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-run-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-openshift-sdn\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.252417 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"sdn-metrics-certs\" (UniqueName: \"kubernetes.io/secret/d5b0f5e6-3d70-4315-b9c5-1403036f517b-sdn-metrics-certs\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.252456 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-run-crio\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-crio\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.252493 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5b0f5e6-3d70-4315-b9c5-1403036f517b-config\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.252521 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-slash\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.252553 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config-mtu-migration\" (UniqueName: \"kubernetes.io/configmap/d5b0f5e6-3d70-4315-b9c5-1403036f517b-config-mtu-migration\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.252582 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-etc-sysconfig\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.252620 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-8nd4t\" (UniqueName: \"kubernetes.io/projected/d5b0f5e6-3d70-4315-b9c5-1403036f517b-kube-api-access-8nd4t\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.252650 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-run-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-kubernetes\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.252681 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-modules\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-modules\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.252721 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-run-ovs\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-ovs\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.252760 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d5b0f5e6-3d70-4315-b9c5-1403036f517b-env-overrides\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.252790 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-cni-bin\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.252826 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-lib-cni-networks-openshift-sdn\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.252859 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-dbus\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.252890 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-cni-conf\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-cni-conf\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.253672 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-etc-sysconfig\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.253731 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-run-netns\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.253774 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-run-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-openshift-sdn\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.260848 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-run-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-kubernetes\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.260923 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-modules\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-modules\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.260970 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-run-ovs\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-ovs\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.261460 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d5b0f5e6-3d70-4315-b9c5-1403036f517b-env-overrides\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.261532 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-cni-bin\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.261586 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-lib-cni-networks-openshift-sdn\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.261631 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-dbus\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.261676 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-cni-conf\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-cni-conf\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.262191 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5b0f5e6-3d70-4315-b9c5-1403036f517b-config\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.262286 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-run-crio\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-crio\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.262386 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-slash\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.262737 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config-mtu-migration\" (UniqueName: \"kubernetes.io/configmap/d5b0f5e6-3d70-4315-b9c5-1403036f517b-config-mtu-migration\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.266620 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"sdn-metrics-certs\" (UniqueName: \"kubernetes.io/secret/d5b0f5e6-3d70-4315-b9c5-1403036f517b-sdn-metrics-certs\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.287788 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nd4t\" (UniqueName: \"kubernetes.io/projected/d5b0f5e6-3d70-4315-b9c5-1403036f517b-kube-api-access-8nd4t\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:05.439885 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:05.440489272Z" level=info msg="Running pod sandbox: openshift-sdn/sdn-66nsp/POD" id=93edfda6-95b5-4f40-956f-9cb8d615d00d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:05.440548651Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:05.488193926Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=93edfda6-95b5-4f40-956f-9cb8d615d00d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:35:05.493239 2927 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5b0f5e6_3d70_4315_b9c5_1403036f517b.slice/crio-878e94265da0a532865f84b39d51b723ec116a6c7849606267000f54329d7b59.scope WatchSource:0}: Error finding container 878e94265da0a532865f84b39d51b723ec116a6c7849606267000f54329d7b59: Status 404 returned error can't find the container with id 878e94265da0a532865f84b39d51b723ec116a6c7849606267000f54329d7b59 Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:05.524097281Z" level=info msg="Ran pod sandbox 878e94265da0a532865f84b39d51b723ec116a6c7849606267000f54329d7b59 with infra container: openshift-sdn/sdn-66nsp/POD" id=93edfda6-95b5-4f40-956f-9cb8d615d00d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:05.526300284Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:30387b6f2e605bb8d018e0c09f9d0275048fc5e7966137fbe6bd4307c88d3fdd" id=51e593e6-59fa-48ea-8cca-0f3728195ee4 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:05.526512929Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:cff39d57d2c205a584f8bf4299b1e125fa1a96ac92f680117ba82c20334754d1,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:30387b6f2e605bb8d018e0c09f9d0275048fc5e7966137fbe6bd4307c88d3fdd],Size_:701241664,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=51e593e6-59fa-48ea-8cca-0f3728195ee4 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:05.527556979Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:30387b6f2e605bb8d018e0c09f9d0275048fc5e7966137fbe6bd4307c88d3fdd" id=8581ec6e-d027-410e-b506-052cf0af7d5f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:05.527739558Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:cff39d57d2c205a584f8bf4299b1e125fa1a96ac92f680117ba82c20334754d1,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:30387b6f2e605bb8d018e0c09f9d0275048fc5e7966137fbe6bd4307c88d3fdd],Size_:701241664,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=8581ec6e-d027-410e-b506-052cf0af7d5f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:05.529374990Z" level=info msg="Creating container: openshift-sdn/sdn-66nsp/sdn" id=c0001e24-f10b-426a-b86b-f58ea9b3f344 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:05.529488124Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:05.715182839Z" level=info msg="Created container c5a34c0e0d97cf3c23b9e200881c870866c518ff644181e7e97dd04f3bf7afe6: openshift-sdn/sdn-66nsp/sdn" id=c0001e24-f10b-426a-b86b-f58ea9b3f344 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:05.717048373Z" level=info msg="Starting container: c5a34c0e0d97cf3c23b9e200881c870866c518ff644181e7e97dd04f3bf7afe6" id=88e3661d-fc8e-4a97-8bf6-7a346d3a0310 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:05.737377485Z" level=info msg="Started container" PID=14572 containerID=c5a34c0e0d97cf3c23b9e200881c870866c518ff644181e7e97dd04f3bf7afe6 description=openshift-sdn/sdn-66nsp/sdn id=88e3661d-fc8e-4a97-8bf6-7a346d3a0310 name=/runtime.v1.RuntimeService/StartContainer sandboxID=878e94265da0a532865f84b39d51b723ec116a6c7849606267000f54329d7b59 Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:05.748570216Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:05.789408805Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:05.789442756Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:05.789462877Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:05.803678027Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=8f33dbdc-81f7-4c6c-8197-e3e05b4add1e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:05.803892789Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=8f33dbdc-81f7-4c6c-8197-e3e05b4add1e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:05.810916879Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=c3934322-d34c-4d50-af0e-241489f8eb58 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:05.811159867Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=c3934322-d34c-4d50-af0e-241489f8eb58 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:05.812594280Z" level=info msg="Creating container: openshift-sdn/sdn-66nsp/kube-rbac-proxy" id=5bfe7c41-dd21-41d9-9663-22f091d0b28d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:05.812709487Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:05.817563632Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:05.817587016Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:05.817606875Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:05.831563976Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:35:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:05.831596423Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:35:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:06.001595 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-sdn/sdn-66nsp" event=&{ID:d5b0f5e6-3d70-4315-b9c5-1403036f517b Type:ContainerStarted Data:c5a34c0e0d97cf3c23b9e200881c870866c518ff644181e7e97dd04f3bf7afe6} Jan 05 09:35:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:06.001627 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-sdn/sdn-66nsp" event=&{ID:d5b0f5e6-3d70-4315-b9c5-1403036f517b Type:ContainerStarted Data:878e94265da0a532865f84b39d51b723ec116a6c7849606267000f54329d7b59} Jan 05 09:35:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:06.176963610Z" level=info msg="Created container 32abadfbb63d6f6d8fd13a849e455f257135842ab04ddc2a58918ae0cef257a5: openshift-sdn/sdn-66nsp/kube-rbac-proxy" id=5bfe7c41-dd21-41d9-9663-22f091d0b28d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:35:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:06.179163156Z" level=info msg="Starting container: 32abadfbb63d6f6d8fd13a849e455f257135842ab04ddc2a58918ae0cef257a5" id=2fc5b828-9b57-455c-85e7-50cc4b347590 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:35:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:06.190006937Z" level=info msg="Started container" PID=14626 containerID=32abadfbb63d6f6d8fd13a849e455f257135842ab04ddc2a58918ae0cef257a5 description=openshift-sdn/sdn-66nsp/kube-rbac-proxy id=2fc5b828-9b57-455c-85e7-50cc4b347590 name=/runtime.v1.RuntimeService/StartContainer sandboxID=878e94265da0a532865f84b39d51b723ec116a6c7849606267000f54329d7b59 Jan 05 09:35:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:07.004671 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-sdn/sdn-66nsp" event=&{ID:d5b0f5e6-3d70-4315-b9c5-1403036f517b Type:ContainerStarted Data:32abadfbb63d6f6d8fd13a849e455f257135842ab04ddc2a58918ae0cef257a5} Jan 05 09:35:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:07.005970 2927 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-sdn/sdn-66nsp" Jan 05 09:35:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:10.476502 2927 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-sdn/sdn-66nsp" podUID=d5b0f5e6-3d70-4315-b9c5-1403036f517b containerName="sdn" probeResult=failure output="" Jan 05 09:35:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:14.648263 2927 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-multus/multus-pbfzz] Jan 05 09:35:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:14.648456 2927 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-multus/multus-pbfzz" podUID=f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7 containerName="kube-multus" containerID="cri-o://da19cbb49f674958c161a6b6f29875eb5d015a5a48e2c25147c41964dd2f2f0c" gracePeriod=10 Jan 05 09:35:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:14.648983554Z" level=info msg="Stopping container: da19cbb49f674958c161a6b6f29875eb5d015a5a48e2c25147c41964dd2f2f0c (timeout: 10s)" id=abbefc2e-fd0b-448b-871b-51b77af12928 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:35:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:15.213035652Z" level=info msg="Stopped container da19cbb49f674958c161a6b6f29875eb5d015a5a48e2c25147c41964dd2f2f0c: openshift-multus/multus-pbfzz/kube-multus" id=abbefc2e-fd0b-448b-871b-51b77af12928 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:35:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:15.213431884Z" level=info msg="Stopping pod sandbox: bcac6a2d9a941220c56968841333b62aa70ad0818ad3c6858d36fafbd613d3b1" id=bb2b6da7-39c4-4333-a7ee-893b56ebe7d0 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:35:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:15.275537312Z" level=info msg="Stopped pod sandbox: bcac6a2d9a941220c56968841333b62aa70ad0818ad3c6858d36fafbd613d3b1" id=bb2b6da7-39c4-4333-a7ee-893b56ebe7d0 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:35:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:15.344647 2927 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-system-cni-dir\") pod \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " Jan 05 09:35:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:15.344702 2927 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tv6rn\" (UniqueName: \"kubernetes.io/projected/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-kube-api-access-tv6rn\") pod \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " Jan 05 09:35:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:15.344730 2927 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-os-release\") pod \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " Jan 05 09:35:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:15.344760 2927 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-multus-cni-dir\") pod \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " Jan 05 09:35:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:15.344792 2927 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-cnibin\") pod \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " Jan 05 09:35:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:15.344824 2927 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-cni-binary-copy\") pod \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\" (UID: \"f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7\") " Jan 05 09:35:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:15.345052 2927 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-os-release" (OuterVolumeSpecName: "os-release") pod "f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7" (UID: "f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7"). InnerVolumeSpecName "os-release". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:35:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:15.345089 2927 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-system-cni-dir" (OuterVolumeSpecName: "system-cni-dir") pod "f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7" (UID: "f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7"). InnerVolumeSpecName "system-cni-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:35:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: W0105 09:35:15.345141 2927 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7/volumes/kubernetes.io~configmap/cni-binary-copy: clearQuota called, but quotas disabled Jan 05 09:35:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:15.345367 2927 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7" (UID: "f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:35:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:15.345400 2927 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-multus-cni-dir" (OuterVolumeSpecName: "multus-cni-dir") pod "f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7" (UID: "f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7"). InnerVolumeSpecName "multus-cni-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:35:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:15.345422 2927 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-cnibin" (OuterVolumeSpecName: "cnibin") pod "f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7" (UID: "f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7"). InnerVolumeSpecName "cnibin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:35:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:15.357034 2927 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-kube-api-access-tv6rn" (OuterVolumeSpecName: "kube-api-access-tv6rn") pod "f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7" (UID: "f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7"). InnerVolumeSpecName "kube-api-access-tv6rn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:35:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:15.445844 2927 reconciler.go:399] "Volume detached for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-system-cni-dir\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:35:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:15.445886 2927 reconciler.go:399] "Volume detached for volume \"kube-api-access-tv6rn\" (UniqueName: \"kubernetes.io/projected/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-kube-api-access-tv6rn\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:35:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:15.445923 2927 reconciler.go:399] "Volume detached for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-os-release\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:35:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:15.445942 2927 reconciler.go:399] "Volume detached for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-multus-cni-dir\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:35:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:15.445962 2927 reconciler.go:399] "Volume detached for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-cnibin\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:35:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:15.445980 2927 reconciler.go:399] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7-cni-binary-copy\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:35:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:15.473647 2927 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-sdn/sdn-66nsp" podUID=d5b0f5e6-3d70-4315-b9c5-1403036f517b containerName="sdn" probeResult=failure output="" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.024344 2927 generic.go:296] "Generic (PLEG): container finished" podID=f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7 containerID="da19cbb49f674958c161a6b6f29875eb5d015a5a48e2c25147c41964dd2f2f0c" exitCode=0 Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.024382 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-pbfzz" event=&{ID:f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7 Type:ContainerDied Data:da19cbb49f674958c161a6b6f29875eb5d015a5a48e2c25147c41964dd2f2f0c} Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.024404 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-pbfzz" event=&{ID:f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7 Type:ContainerDied Data:bcac6a2d9a941220c56968841333b62aa70ad0818ad3c6858d36fafbd613d3b1} Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.024421 2927 scope.go:115] "RemoveContainer" containerID="da19cbb49f674958c161a6b6f29875eb5d015a5a48e2c25147c41964dd2f2f0c" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:16.027883395Z" level=info msg="Removing container: da19cbb49f674958c161a6b6f29875eb5d015a5a48e2c25147c41964dd2f2f0c" id=2490b590-a886-4277-a7cd-2294b978863e name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.051113 2927 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-multus/multus-pbfzz] Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.056279 2927 kubelet.go:2129] "SyncLoop REMOVE" source="api" pods=[openshift-multus/multus-pbfzz] Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:16.060934143Z" level=info msg="Removed container da19cbb49f674958c161a6b6f29875eb5d015a5a48e2c25147c41964dd2f2f0c: openshift-multus/multus-pbfzz/kube-multus" id=2490b590-a886-4277-a7cd-2294b978863e name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.061101 2927 scope.go:115] "RemoveContainer" containerID="da19cbb49f674958c161a6b6f29875eb5d015a5a48e2c25147c41964dd2f2f0c" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:35:16.061407 2927 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da19cbb49f674958c161a6b6f29875eb5d015a5a48e2c25147c41964dd2f2f0c\": container with ID starting with da19cbb49f674958c161a6b6f29875eb5d015a5a48e2c25147c41964dd2f2f0c not found: ID does not exist" containerID="da19cbb49f674958c161a6b6f29875eb5d015a5a48e2c25147c41964dd2f2f0c" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.061450 2927 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:da19cbb49f674958c161a6b6f29875eb5d015a5a48e2c25147c41964dd2f2f0c} err="failed to get container status \"da19cbb49f674958c161a6b6f29875eb5d015a5a48e2c25147c41964dd2f2f0c\": rpc error: code = NotFound desc = could not find container \"da19cbb49f674958c161a6b6f29875eb5d015a5a48e2c25147c41964dd2f2f0c\": container with ID starting with da19cbb49f674958c161a6b6f29875eb5d015a5a48e2c25147c41964dd2f2f0c not found: ID does not exist" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.065758 2927 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-multus/multus-4xhw7] Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.065798 2927 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:35:16.065874 2927 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7" containerName="kube-multus" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.065888 2927 state_mem.go:107] "Deleted CPUSet assignment" podUID="f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7" containerName="kube-multus" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.065990 2927 memory_manager.go:345] "RemoveStaleState removing state" podUID="f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7" containerName="kube-multus" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.150195 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zf2qr\" (UniqueName: \"kubernetes.io/projected/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-kube-api-access-zf2qr\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.150307 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-os-release\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.150353 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-system-cni-dir\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.150413 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-multus-cni-dir\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.150459 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-cnibin\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.150510 2927 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-cni-binary-copy\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.251424 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-system-cni-dir\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.251478 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-multus-cni-dir\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.251513 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-cnibin\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.251561 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-cni-binary-copy\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.251615 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-zf2qr\" (UniqueName: \"kubernetes.io/projected/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-kube-api-access-zf2qr\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.251646 2927 reconciler.go:269] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-os-release\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.251746 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-os-release\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.251814 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-system-cni-dir\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.251876 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-multus-cni-dir\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.251954 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-cnibin\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.252525 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-cni-binary-copy\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.270330 2927 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-zf2qr\" (UniqueName: \"kubernetes.io/projected/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-kube-api-access-zf2qr\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.384283 2927 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-4xhw7" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:16.385005181Z" level=info msg="Running pod sandbox: openshift-multus/multus-4xhw7/POD" id=9bbf395e-b612-4c73-b439-b401d7b147bd name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:16.385065758Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:16.491597047Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=9bbf395e-b612-4c73-b439-b401d7b147bd name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:16.517072142Z" level=info msg="Ran pod sandbox 9901245459f8da67113c56c8790e3238fd9fb07f4a0ba858fa646867a4aad700 with infra container: openshift-multus/multus-4xhw7/POD" id=9bbf395e-b612-4c73-b439-b401d7b147bd name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:16.527472191Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=42c9d5b4-912f-4851-ad11-7aa4ed151447 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:16.527733509Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=42c9d5b4-912f-4851-ad11-7aa4ed151447 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:16.528770599Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=629aea81-3a98-40b6-b330-e9b488ce64e0 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:16.528993396Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=629aea81-3a98-40b6-b330-e9b488ce64e0 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:16.530120495Z" level=info msg="Creating container: openshift-multus/multus-4xhw7/kube-multus" id=430238a9-d21b-4255-b735-3f0f22008bc6 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:16.530242065Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:16.695022567Z" level=info msg="Created container 1eb0e14ebebbaad840a152559b86d7eee7f1e2e2469fb426f588f96a49c735ba: openshift-multus/multus-4xhw7/kube-multus" id=430238a9-d21b-4255-b735-3f0f22008bc6 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:16.695586240Z" level=info msg="Starting container: 1eb0e14ebebbaad840a152559b86d7eee7f1e2e2469fb426f588f96a49c735ba" id=cb83e6c2-52ec-4d0d-94a9-b306e24862cf name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:16.703896445Z" level=info msg="Started container" PID=14839 containerID=1eb0e14ebebbaad840a152559b86d7eee7f1e2e2469fb426f588f96a49c735ba description=openshift-multus/multus-4xhw7/kube-multus id=cb83e6c2-52ec-4d0d-94a9-b306e24862cf name=/runtime.v1.RuntimeService/StartContainer sandboxID=9901245459f8da67113c56c8790e3238fd9fb07f4a0ba858fa646867a4aad700 Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:16.716256332Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_4f636e92-8068-419f-94b0-f619ab850094\"" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:16.728657783Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:16.728687153Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:16.754475368Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/multus\"" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:16.767273966Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:16.767304751Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:16.767323816Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_4f636e92-8068-419f-94b0-f619ab850094\"" Jan 05 09:35:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:16.945041 2927 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7 path="/var/lib/kubelet/pods/f907add9-a2cf-4fc7-b7f8-f283bc1c2dd7/volumes" Jan 05 09:35:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:17.027178 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4xhw7" event=&{ID:84f9cd8a-5e58-4aff-aedd-906ed37dc97d Type:ContainerStarted Data:1eb0e14ebebbaad840a152559b86d7eee7f1e2e2469fb426f588f96a49c735ba} Jan 05 09:35:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:17.027267 2927 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4xhw7" event=&{ID:84f9cd8a-5e58-4aff-aedd-906ed37dc97d Type:ContainerStarted Data:9901245459f8da67113c56c8790e3238fd9fb07f4a0ba858fa646867a4aad700} Jan 05 09:35:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:18.371417694Z" level=warning msg="Stdout copy error: read /dev/ptmx: input/output error" Jan 05 09:35:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:20.477261 2927 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-sdn/sdn-66nsp" podUID=d5b0f5e6-3d70-4315-b9c5-1403036f517b containerName="sdn" probeResult=failure output="" Jan 05 09:35:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:25.474048 2927 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-sdn/sdn-66nsp" podUID=d5b0f5e6-3d70-4315-b9c5-1403036f517b containerName="sdn" probeResult=failure output="" Jan 05 09:35:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:28.754011106Z" level=info msg="Stopping pod sandbox: fd9deb131d1f752a07a6704193bcd1cf1dab19ff2e6439be2d911f94c338da51" id=5cb1e166-476b-4dbe-b1f5-3979e6f685e7 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:35:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:28.754057149Z" level=info msg="Stopped pod sandbox (already stopped): fd9deb131d1f752a07a6704193bcd1cf1dab19ff2e6439be2d911f94c338da51" id=5cb1e166-476b-4dbe-b1f5-3979e6f685e7 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:35:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:28.754306323Z" level=info msg="Removing pod sandbox: fd9deb131d1f752a07a6704193bcd1cf1dab19ff2e6439be2d911f94c338da51" id=b9bdee13-28c3-47c3-9663-9902fd9bf5a4 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:35:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:28.764132646Z" level=info msg="Removed pod sandbox: fd9deb131d1f752a07a6704193bcd1cf1dab19ff2e6439be2d911f94c338da51" id=b9bdee13-28c3-47c3-9663-9902fd9bf5a4 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:35:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:28.764438373Z" level=info msg="Stopping pod sandbox: bcac6a2d9a941220c56968841333b62aa70ad0818ad3c6858d36fafbd613d3b1" id=5fd9a2e0-f088-4546-8e8c-1f76e1b00f98 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:35:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:28.764467910Z" level=info msg="Stopped pod sandbox (already stopped): bcac6a2d9a941220c56968841333b62aa70ad0818ad3c6858d36fafbd613d3b1" id=5fd9a2e0-f088-4546-8e8c-1f76e1b00f98 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:35:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:28.764739604Z" level=info msg="Removing pod sandbox: bcac6a2d9a941220c56968841333b62aa70ad0818ad3c6858d36fafbd613d3b1" id=f19adfe4-89c0-4fa4-b863-50ce1ddb988c name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:35:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:28.772801813Z" level=info msg="Removed pod sandbox: bcac6a2d9a941220c56968841333b62aa70ad0818ad3c6858d36fafbd613d3b1" id=f19adfe4-89c0-4fa4-b863-50ce1ddb988c name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:35:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: E0105 09:35:28.774395 2927 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fed50c900e7456872d20306aa0e63b13927facd5fb4913292c3036a63611a79\": container with ID starting with 4fed50c900e7456872d20306aa0e63b13927facd5fb4913292c3036a63611a79 not found: ID does not exist" containerID="4fed50c900e7456872d20306aa0e63b13927facd5fb4913292c3036a63611a79" Jan 05 09:35:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:28.774435 2927 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="4fed50c900e7456872d20306aa0e63b13927facd5fb4913292c3036a63611a79" err="rpc error: code = NotFound desc = could not find container \"4fed50c900e7456872d20306aa0e63b13927facd5fb4913292c3036a63611a79\": container with ID starting with 4fed50c900e7456872d20306aa0e63b13927facd5fb4913292c3036a63611a79 not found: ID does not exist" Jan 05 09:35:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:30.475476 2927 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-sdn/sdn-66nsp" podUID=d5b0f5e6-3d70-4315-b9c5-1403036f517b containerName="sdn" probeResult=failure output="" Jan 05 09:35:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:35.517803 2927 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-sdn/sdn-66nsp" podUID=d5b0f5e6-3d70-4315-b9c5-1403036f517b containerName="sdn" probeResult=failure output="" Jan 05 09:35:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:35:36.306682688Z" level=error msg="Failed to cleanup (probably retrying): failed to destroy network for pod sandbox k8s_thanos-querier-6f4f5fb7cf-rdm9l_openshift-monitoring_7c63c9be-d1ad-44b6-8b33-09af3ee314af_0(f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3): error removing pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/7c63c9be-d1ad-44b6-8b33-09af3ee314af]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" Jan 05 09:35:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:40.474342 2927 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-sdn/sdn-66nsp" podUID=d5b0f5e6-3d70-4315-b9c5-1403036f517b containerName="sdn" probeResult=failure output="" Jan 05 09:35:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:45.475507 2927 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-sdn/sdn-66nsp" podUID=d5b0f5e6-3d70-4315-b9c5-1403036f517b containerName="sdn" probeResult=failure output="" Jan 05 09:35:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:50.481165 2927 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-sdn/sdn-66nsp" podUID=d5b0f5e6-3d70-4315-b9c5-1403036f517b containerName="sdn" probeResult=failure output="" Jan 05 09:35:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:35:55.476997 2927 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-sdn/sdn-66nsp" podUID=d5b0f5e6-3d70-4315-b9c5-1403036f517b containerName="sdn" probeResult=failure output="" Jan 05 09:36:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:36:00.480303 2927 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-sdn/sdn-66nsp" podUID=d5b0f5e6-3d70-4315-b9c5-1403036f517b containerName="sdn" probeResult=failure output="" Jan 05 09:36:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:36:05.501237 2927 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-sdn/sdn-66nsp" podUID=d5b0f5e6-3d70-4315-b9c5-1403036f517b containerName="sdn" probeResult=failure output="" Jan 05 09:36:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:36:10.481234 2927 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-sdn/sdn-66nsp" podUID=d5b0f5e6-3d70-4315-b9c5-1403036f517b containerName="sdn" probeResult=failure output="" Jan 05 09:36:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:36:15.473520 2927 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-sdn/sdn-66nsp" podUID=d5b0f5e6-3d70-4315-b9c5-1403036f517b containerName="sdn" probeResult=failure output="" Jan 05 09:36:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:36:17.977795 2927 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-sdn/sdn-66nsp" podUID=d5b0f5e6-3d70-4315-b9c5-1403036f517b containerName="sdn" probeResult=failure output="" Jan 05 09:36:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Stopping Kubernetes Kubelet... Jan 05 09:36:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:36:18.525695 2927 plugin_watcher.go:215] "Removing socket path from desired state cache" path="/var/lib/kubelet/plugins_registry/diskplugin.csi.alibabacloud.com-reg.sock" Jan 05 09:36:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2927]: I0105 09:36:18.525751 2927 dynamic_cafile_content.go:171] "Shutting down controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 05 09:36:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: kubelet.service: Succeeded. Jan 05 09:36:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Stopped Kubernetes Kubelet. Jan 05 09:36:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: kubelet.service: Consumed 30.768s CPU time Jan 05 09:36:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:36:19.571101443Z" level=info msg="cleanup sandbox network" Jan 05 09:36:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:36:19.575148812Z" level=info msg="Got pod network &{Name:thanos-querier-6f4f5fb7cf-rdm9l Namespace:openshift-monitoring ID:f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3 UID:7c63c9be-d1ad-44b6-8b33-09af3ee314af NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:36:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:36:19.575509524Z" level=info msg="Deleting pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:37:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Stopping Container Runtime Interface for OCI (CRI-O)... Jan 05 09:37:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:37:04.518513929Z" level=error msg="Failed to update container state for 81d748520620965f53fabe92f19a5bb3c5d4e2f876e096dcb82692d2f2f8f610: `/usr/bin/runc --root /run/runc --systemd-cgroup state 81d748520620965f53fabe92f19a5bb3c5d4e2f876e096dcb82692d2f2f8f610` failed: : signal: terminated" Jan 05 09:37:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2879]: time="2023-01-05 09:37:04.519012958Z" level=error msg="Failed to cleanup (probably retrying): failed to destroy network for pod sandbox k8s_thanos-querier-6f4f5fb7cf-rdm9l_openshift-monitoring_7c63c9be-d1ad-44b6-8b33-09af3ee314af_0(f0db612f6c6539708f21ddc48dbe81ec6e4499759c71108c5edfdd4043cb71d3): error removing pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): netplugin failed with no error message: signal: terminated" Jan 05 09:37:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: crio.service: Succeeded. Jan 05 09:37:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Stopped Container Runtime Interface for OCI (CRI-O). Jan 05 09:37:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: crio.service: Consumed 38.774s CPU time -- Reboot -- Jan 05 17:37:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Starting Container Runtime Interface for OCI (CRI-O)... Jan 05 17:37:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 17:37:38.653975598Z" level=info msg="Starting CRI-O, version: 1.25.1-5.rhaos4.12.git6005903.el8, git: unknown(clean)" Jan 05 17:37:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 17:37:38.654348219Z" level=info msg="Node configuration value for hugetlb cgroup is true" Jan 05 17:37:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 17:37:38.654360982Z" level=info msg="Node configuration value for pid cgroup is true" Jan 05 17:37:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 17:37:38.654414252Z" level=info msg="Node configuration value for memoryswap cgroup is true" Jan 05 17:37:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 17:37:38.654422097Z" level=info msg="Node configuration value for cgroup v2 is false" Jan 05 17:37:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 17:37:38.663113582Z" level=info msg="Node configuration value for systemd CollectMode is true" Jan 05 17:37:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 17:37:38.669477650Z" level=info msg="Node configuration value for systemd AllowedCPUs is true" Jan 05 17:37:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 17:37:38.673280632Z" level=info msg="Using default capabilities: CAP_CHOWN, CAP_DAC_OVERRIDE, CAP_FSETID, CAP_FOWNER, CAP_SETGID, CAP_SETUID, CAP_SETPCAP, CAP_NET_BIND_SERVICE, CAP_KILL" Jan 05 17:37:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 17:37:38.749645571Z" level=info msg="Checkpoint/restore support disabled" Jan 05 17:37:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 17:37:38.750024012Z" level=info msg="Using seccomp default profile when unspecified: true" Jan 05 17:37:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 17:37:38.750113014Z" level=info msg="Using the internal default seccomp profile" Jan 05 17:37:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 17:37:38.750125561Z" level=info msg="AppArmor is disabled by the system or at CRI-O build-time" Jan 05 17:37:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 17:37:38.750147728Z" level=info msg="No blockio config file specified, blockio not configured" Jan 05 17:37:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 17:37:38.750156309Z" level=info msg="RDT not available in the host system" Jan 05 17:37:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 17:37:38.755515994Z" level=info msg="Conmon does support the --sync option" Jan 05 17:37:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 17:37:38.755580130Z" level=info msg="Conmon does support the --log-global-size-max option" Jan 05 17:37:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 17:37:38.759230627Z" level=info msg="Conmon does support the --sync option" Jan 05 17:37:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 17:37:38.759248613Z" level=info msg="Conmon does support the --log-global-size-max option" Jan 05 17:37:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 17:37:38.867294620Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 17:37:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 17:37:38.867326663Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 17:37:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 17:37:38.921206692Z" level=warning msg="Could not restore sandbox 04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0: failed to Statfs \"/var/run/netns/358f9ab6-2e3e-47df-b3f3-eb6ac7120ea1\": no such file or directory" Jan 05 17:37:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 17:37:38.934815600Z" level=warning msg="Deleting all containers under sandbox 04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0 since it could not be restored" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.468696026Z" level=warning msg="Could not restore sandbox 318b2d4d118085d70f305dfbfb7875ec65ba1a3f9539d6e1f29f152cb2d8af82: failed to Statfs \"/var/run/netns/d458dea1-0719-4a13-bc09-29a80e493860\": no such file or directory" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.486444826Z" level=warning msg="Deleting all containers under sandbox 318b2d4d118085d70f305dfbfb7875ec65ba1a3f9539d6e1f29f152cb2d8af82 since it could not be restored" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.512452327Z" level=warning msg="Could not restore sandbox 878e94265da0a532865f84b39d51b723ec116a6c7849606267000f54329d7b59: failed to Statfs \"/var/run/netns/49f088d3-f3b6-4ea2-8486-57b900eca477\": no such file or directory" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.520170752Z" level=warning msg="Deleting all containers under sandbox 878e94265da0a532865f84b39d51b723ec116a6c7849606267000f54329d7b59 since it could not be restored" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.540183678Z" level=warning msg="Could not restore sandbox 9901245459f8da67113c56c8790e3238fd9fb07f4a0ba858fa646867a4aad700: failed to Statfs \"/var/run/netns/af1d963b-4d9f-4f4a-b058-2d29008ae103\": no such file or directory" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.550149226Z" level=warning msg="Deleting all containers under sandbox 9901245459f8da67113c56c8790e3238fd9fb07f4a0ba858fa646867a4aad700 since it could not be restored" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.563204433Z" level=warning msg="Could not restore sandbox a22e1d52a296a877749d67bb8347e63d619ee2475777ed73628f6d33736e68d4: failed to Statfs \"/var/run/netns/05ac99ed-47b8-4ac4-b846-f8535b6ff651\": no such file or directory" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.571883147Z" level=warning msg="Deleting all containers under sandbox a22e1d52a296a877749d67bb8347e63d619ee2475777ed73628f6d33736e68d4 since it could not be restored" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.579001097Z" level=warning msg="Could not restore sandbox fe3e7b123db46a0b88eddb5d106d43267fc1d0c1707470c0568f2b211ee6ba55: failed to Statfs \"/var/run/netns/d1d94e66-dbb6-4156-bf1f-528844201625\": no such file or directory" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.585388280Z" level=warning msg="Deleting all containers under sandbox fe3e7b123db46a0b88eddb5d106d43267fc1d0c1707470c0568f2b211ee6ba55 since it could not be restored" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.590838931Z" level=warning msg="Could not restore sandbox 90d2b912f09210aef9996180fedf7e65296d392716fb68985112976a13c8d29b: failed to Statfs \"/var/run/netns/58506b8e-16a1-4c50-8f8b-a8b3adf8da54\": no such file or directory" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.598808313Z" level=warning msg="Deleting all containers under sandbox 90d2b912f09210aef9996180fedf7e65296d392716fb68985112976a13c8d29b since it could not be restored" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.605736380Z" level=warning msg="Could not restore sandbox 8a313b3accfad3269666173b7b1c660df2741df597e595c3c47a5ca7eb07e228: failed to Statfs \"/var/run/netns/c226eea0-515f-4498-858c-3c6f29b4b2db\": no such file or directory" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.612643677Z" level=warning msg="Deleting all containers under sandbox 8a313b3accfad3269666173b7b1c660df2741df597e595c3c47a5ca7eb07e228 since it could not be restored" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.617978742Z" level=warning msg="Could not restore sandbox e0a041e795d1586c6307ac859c48d38830f232aa0328f519a4f2db991d42e38a: failed to Statfs \"/var/run/netns/c196c4ae-f75c-4a2d-8f59-85c45ced434c\": no such file or directory" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.623505350Z" level=warning msg="Deleting all containers under sandbox e0a041e795d1586c6307ac859c48d38830f232aa0328f519a4f2db991d42e38a since it could not be restored" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.661252647Z" level=warning msg="Could not restore sandbox 0a799b0f1916c3d4fc548b401a6c06c1907860552ea55059042c2484d4fa68dc: failed to Statfs \"/var/run/netns/1775d78f-2f22-4c44-a337-b141084391ef\": no such file or directory" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.668226927Z" level=warning msg="Deleting all containers under sandbox 0a799b0f1916c3d4fc548b401a6c06c1907860552ea55059042c2484d4fa68dc since it could not be restored" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.678229889Z" level=warning msg="Could not restore sandbox 23c255a61eb7c200b97ccd2916bb330a59475bd2d3e145171f1e9f94a9c27e2e: failed to Statfs \"/var/run/netns/804277e1-aeb2-45f0-b45d-ec85b6d5c5ce\": no such file or directory" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.684252455Z" level=warning msg="Deleting all containers under sandbox 23c255a61eb7c200b97ccd2916bb330a59475bd2d3e145171f1e9f94a9c27e2e since it could not be restored" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.714694689Z" level=warning msg="Could not restore sandbox f16d1d77b76cea9d449631cf97d4fe890b2fee1b5dc69a019689afaeec8f7b05: failed to Statfs \"/var/run/netns/3efcca22-5def-4949-9634-189b4370a742\": no such file or directory" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.720974439Z" level=warning msg="Deleting all containers under sandbox f16d1d77b76cea9d449631cf97d4fe890b2fee1b5dc69a019689afaeec8f7b05 since it could not be restored" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.726906879Z" level=warning msg="Could not restore sandbox 592c0fe68eaae38ffff4788882a53055cedb618ddc573956838720a91f262ddf: failed to Statfs \"/var/run/netns/44c3e3ce-0d3a-4e35-9d33-048ce0d923c0\": no such file or directory" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.733285759Z" level=warning msg="Deleting all containers under sandbox 592c0fe68eaae38ffff4788882a53055cedb618ddc573956838720a91f262ddf since it could not be restored" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.752537154Z" level=warning msg="Could not restore sandbox 68d9bc9c6a9bcf27e7bcf19bef0f4b3c6707eefbb0ba0e388f51a8e663a95ad7: failed to Statfs \"/var/run/netns/b8f70c62-7b79-4f76-b3e5-1851b2785ffa\": no such file or directory" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.759251555Z" level=warning msg="Deleting all containers under sandbox 68d9bc9c6a9bcf27e7bcf19bef0f4b3c6707eefbb0ba0e388f51a8e663a95ad7 since it could not be restored" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.767447770Z" level=warning msg="Could not restore sandbox 238eef320be8a6fa61ec83ea4bf9cdc139cde1247f51dbe0bd0040a8e9c70088: failed to Statfs \"/var/run/netns/8e512d8a-b52a-45da-a484-9148509a8cea\": no such file or directory" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.773891847Z" level=warning msg="Deleting all containers under sandbox 238eef320be8a6fa61ec83ea4bf9cdc139cde1247f51dbe0bd0040a8e9c70088 since it could not be restored" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.784154566Z" level=warning msg="Could not restore sandbox 9c318dae338ceca029f5b8d71eed7d2265d6188623995ec40e8d80da65a8d993: failed to Statfs \"/var/run/netns/8598084b-1708-4476-81ee-e77d49e075ea\": no such file or directory" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.790386933Z" level=warning msg="Deleting all containers under sandbox 9c318dae338ceca029f5b8d71eed7d2265d6188623995ec40e8d80da65a8d993 since it could not be restored" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.808554014Z" level=warning msg="Could not restore sandbox 2de3677cadab589863eceefc75ced2bea032cf20d59c2b31ab33b3b6e1993538: failed to Statfs \"/var/run/netns/a33e6009-940a-4247-8870-bfa8301b0f21\": no such file or directory" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.814080609Z" level=warning msg="Deleting all containers under sandbox 2de3677cadab589863eceefc75ced2bea032cf20d59c2b31ab33b3b6e1993538 since it could not be restored" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.832474304Z" level=warning msg="Could not restore sandbox 8ea5e7e9fb15db146766a6199273736a81c0e10f5714578c6922f99837137af4: failed to Statfs \"/var/run/netns/9b9707ac-7d19-44b0-95c3-002e12244dde\": no such file or directory" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.841031282Z" level=warning msg="Deleting all containers under sandbox 8ea5e7e9fb15db146766a6199273736a81c0e10f5714578c6922f99837137af4 since it could not be restored" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.877490270Z" level=warning msg="Could not restore sandbox c42a602a42d67c9af5ab623c3df747e5144697eb541c212fccf9ff7c0f6542d7: failed to Statfs \"/var/run/netns/d6a26d6a-1b05-4d09-ab18-288b0d18d0e6\": no such file or directory" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.883553912Z" level=warning msg="Deleting all containers under sandbox c42a602a42d67c9af5ab623c3df747e5144697eb541c212fccf9ff7c0f6542d7 since it could not be restored" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.894807960Z" level=warning msg="Could not restore sandbox f1e02d518b5e62edc187ed7cc730fab2279f2eb5384e3f3ffd25d3a014b31d4b: failed to Statfs \"/var/run/netns/04b13e36-0acc-4f75-9752-0ebb73acd4ae\": no such file or directory" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.900241471Z" level=warning msg="Deleting all containers under sandbox f1e02d518b5e62edc187ed7cc730fab2279f2eb5384e3f3ffd25d3a014b31d4b since it could not be restored" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.914797567Z" level=warning msg="Could not restore sandbox de5e12e9744f461b9b6c0146346e5117bb31a1f6538299e5dccde6d8250c4268: failed to Statfs \"/var/run/netns/6d71acfd-733e-4abe-9500-a64746e72ca6\": no such file or directory" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.920050513Z" level=warning msg="Deleting all containers under sandbox de5e12e9744f461b9b6c0146346e5117bb31a1f6538299e5dccde6d8250c4268 since it could not be restored" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.966149556Z" level=warning msg="Could not restore sandbox f226c6f82e17c96c33e4b6a3a2d81c74b2301377f9323d0a64195f4af1cc5679: failed to Statfs \"/var/run/netns/cfdfca99-9582-448d-b3cd-a75ebd263536\": no such file or directory" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.971238460Z" level=warning msg="Deleting all containers under sandbox f226c6f82e17c96c33e4b6a3a2d81c74b2301377f9323d0a64195f4af1cc5679 since it could not be restored" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.976205686Z" level=warning msg="Error encountered when checking whether cri-o should wipe containers: open /var/run/crio/version: no such file or directory" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.976229288Z" level=info msg="cleanup sandbox network" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.977006356Z" level=info msg="Successfully cleaned up network for pod 0a799b0f1916c3d4fc548b401a6c06c1907860552ea55059042c2484d4fa68dc" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.977020746Z" level=info msg="cleanup sandbox network" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.977030093Z" level=info msg="Successfully cleaned up network for pod 878e94265da0a532865f84b39d51b723ec116a6c7849606267000f54329d7b59" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.977040397Z" level=info msg="cleanup sandbox network" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.979976833Z" level=info msg="Got pod network &{Name:prometheus-k8s-0 Namespace:openshift-monitoring ID:04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0 UID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.980363248Z" level=info msg="Serving metrics on :9537 via HTTP" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:39.981136727Z" level=info msg="Deleting pod openshift-monitoring_prometheus-k8s-0 from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Started Container Runtime Interface for OCI (CRI-O). Jan 05 09:37:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Starting Kubernetes Kubelet... Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: Flag --container-runtime has been deprecated, will be removed in 1.27 as the only valid value is 'remote' Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: Flag --cloud-provider has been deprecated, will be removed in 1.25 or later, in favor of removing cloud provider code from Kubelet. Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: Flag --provider-id has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.27. Image garbage collector will get sandbox image information from CRI. Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.352076 2588 server.go:200] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: Flag --container-runtime has been deprecated, will be removed in 1.27 as the only valid value is 'remote' Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: Flag --cloud-provider has been deprecated, will be removed in 1.25 or later, in favor of removing cloud provider code from Kubelet. Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: Flag --provider-id has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.27. Image garbage collector will get sandbox image information from CRI. Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354128 2588 flags.go:64] FLAG: --add-dir-header="false" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354146 2588 flags.go:64] FLAG: --address="0.0.0.0" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354154 2588 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354165 2588 flags.go:64] FLAG: --alsologtostderr="false" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354171 2588 flags.go:64] FLAG: --anonymous-auth="true" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354177 2588 flags.go:64] FLAG: --application-metrics-count-limit="100" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354183 2588 flags.go:64] FLAG: --authentication-token-webhook="false" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354188 2588 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354196 2588 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354205 2588 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354211 2588 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354216 2588 flags.go:64] FLAG: --azure-container-registry-config="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354222 2588 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354228 2588 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354234 2588 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354240 2588 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354255 2588 flags.go:64] FLAG: --cgroup-root="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354260 2588 flags.go:64] FLAG: --cgroups-per-qos="true" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354280 2588 flags.go:64] FLAG: --client-ca-file="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354285 2588 flags.go:64] FLAG: --cloud-config="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354290 2588 flags.go:64] FLAG: --cloud-provider="external" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354296 2588 flags.go:64] FLAG: --cluster-dns="[]" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354302 2588 flags.go:64] FLAG: --cluster-domain="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354308 2588 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354315 2588 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354321 2588 flags.go:64] FLAG: --container-log-max-files="5" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354328 2588 flags.go:64] FLAG: --container-log-max-size="10Mi" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354334 2588 flags.go:64] FLAG: --container-runtime="remote" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354339 2588 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354344 2588 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354350 2588 flags.go:64] FLAG: --containerd-namespace="k8s.io" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354356 2588 flags.go:64] FLAG: --contention-profiling="false" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354361 2588 flags.go:64] FLAG: --cpu-cfs-quota="true" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354367 2588 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354372 2588 flags.go:64] FLAG: --cpu-manager-policy="none" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354380 2588 flags.go:64] FLAG: --cpu-manager-policy-options="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354388 2588 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354394 2588 flags.go:64] FLAG: --enable-controller-attach-detach="true" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354399 2588 flags.go:64] FLAG: --enable-debugging-handlers="true" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354404 2588 flags.go:64] FLAG: --enable-load-reader="false" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354409 2588 flags.go:64] FLAG: --enable-server="true" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354414 2588 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354424 2588 flags.go:64] FLAG: --event-burst="10" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354430 2588 flags.go:64] FLAG: --event-qps="5" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354435 2588 flags.go:64] FLAG: --event-storage-age-limit="default=0" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354441 2588 flags.go:64] FLAG: --event-storage-event-limit="default=0" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354447 2588 flags.go:64] FLAG: --eviction-hard="imagefs.available<15%,memory.available<100Mi,nodefs.available<10%,nodefs.inodesFree<5%" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354462 2588 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354468 2588 flags.go:64] FLAG: --eviction-minimum-reclaim="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354473 2588 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354479 2588 flags.go:64] FLAG: --eviction-soft="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354484 2588 flags.go:64] FLAG: --eviction-soft-grace-period="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354489 2588 flags.go:64] FLAG: --exit-on-lock-contention="false" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354495 2588 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354504 2588 flags.go:64] FLAG: --experimental-mounter-path="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354509 2588 flags.go:64] FLAG: --fail-swap-on="true" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354515 2588 flags.go:64] FLAG: --feature-gates="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354521 2588 flags.go:64] FLAG: --file-check-frequency="20s" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354526 2588 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354532 2588 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354538 2588 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354544 2588 flags.go:64] FLAG: --healthz-port="10248" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354551 2588 flags.go:64] FLAG: --help="false" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354557 2588 flags.go:64] FLAG: --hostname-override="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354562 2588 flags.go:64] FLAG: --housekeeping-interval="10s" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354568 2588 flags.go:64] FLAG: --http-check-frequency="20s" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354573 2588 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354578 2588 flags.go:64] FLAG: --image-credential-provider-config="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354583 2588 flags.go:64] FLAG: --image-gc-high-threshold="85" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354588 2588 flags.go:64] FLAG: --image-gc-low-threshold="80" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354593 2588 flags.go:64] FLAG: --image-service-endpoint="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354598 2588 flags.go:64] FLAG: --iptables-drop-bit="15" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354603 2588 flags.go:64] FLAG: --iptables-masquerade-bit="14" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354613 2588 flags.go:64] FLAG: --keep-terminated-pod-volumes="false" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354619 2588 flags.go:64] FLAG: --kernel-memcg-notification="false" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354627 2588 flags.go:64] FLAG: --kube-api-burst="10" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354632 2588 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354639 2588 flags.go:64] FLAG: --kube-api-qps="5" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354644 2588 flags.go:64] FLAG: --kube-reserved="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354650 2588 flags.go:64] FLAG: --kube-reserved-cgroup="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354655 2588 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354680 2588 flags.go:64] FLAG: --kubelet-cgroups="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354685 2588 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354691 2588 flags.go:64] FLAG: --lock-file="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354696 2588 flags.go:64] FLAG: --log-backtrace-at=":0" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354703 2588 flags.go:64] FLAG: --log-cadvisor-usage="false" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354708 2588 flags.go:64] FLAG: --log-dir="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354713 2588 flags.go:64] FLAG: --log-file="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354718 2588 flags.go:64] FLAG: --log-file-max-size="1800" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354724 2588 flags.go:64] FLAG: --log-flush-frequency="5s" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354729 2588 flags.go:64] FLAG: --log-json-info-buffer-size="0" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354738 2588 flags.go:64] FLAG: --log-json-split-stream="false" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354744 2588 flags.go:64] FLAG: --logging-format="text" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354756 2588 flags.go:64] FLAG: --logtostderr="true" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354762 2588 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354768 2588 flags.go:64] FLAG: --make-iptables-util-chains="true" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354773 2588 flags.go:64] FLAG: --manifest-url="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354778 2588 flags.go:64] FLAG: --manifest-url-header="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354786 2588 flags.go:64] FLAG: --master-service-namespace="default" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354791 2588 flags.go:64] FLAG: --max-housekeeping-interval="15s" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354796 2588 flags.go:64] FLAG: --max-open-files="1000000" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354803 2588 flags.go:64] FLAG: --max-pods="110" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354808 2588 flags.go:64] FLAG: --maximum-dead-containers="-1" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354813 2588 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354819 2588 flags.go:64] FLAG: --memory-manager-policy="None" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354823 2588 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354829 2588 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354835 2588 flags.go:64] FLAG: --node-ip="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354839 2588 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/worker=,node.openshift.io/os_id=rhcos" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354851 2588 flags.go:64] FLAG: --node-status-max-images="50" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354856 2588 flags.go:64] FLAG: --node-status-update-frequency="10s" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354862 2588 flags.go:64] FLAG: --one-output="false" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354867 2588 flags.go:64] FLAG: --oom-score-adj="-999" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354873 2588 flags.go:64] FLAG: --pod-cidr="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354878 2588 flags.go:64] FLAG: --pod-infra-container-image="registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354886 2588 flags.go:64] FLAG: --pod-manifest-path="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354892 2588 flags.go:64] FLAG: --pod-max-pids="-1" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354897 2588 flags.go:64] FLAG: --pods-per-core="0" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354902 2588 flags.go:64] FLAG: --port="10250" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354907 2588 flags.go:64] FLAG: --protect-kernel-defaults="false" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354913 2588 flags.go:64] FLAG: --provider-id="alicloud://us-east-1.i-0xif5dui7khr4rm1wp68" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354918 2588 flags.go:64] FLAG: --qos-reserved="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354924 2588 flags.go:64] FLAG: --read-only-port="10255" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354929 2588 flags.go:64] FLAG: --register-node="true" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354934 2588 flags.go:64] FLAG: --register-schedulable="true" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354941 2588 flags.go:64] FLAG: --register-with-taints="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354948 2588 flags.go:64] FLAG: --registry-burst="10" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354953 2588 flags.go:64] FLAG: --registry-qps="5" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354959 2588 flags.go:64] FLAG: --reserved-cpus="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354964 2588 flags.go:64] FLAG: --reserved-memory="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354970 2588 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354976 2588 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354983 2588 flags.go:64] FLAG: --rotate-certificates="false" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354988 2588 flags.go:64] FLAG: --rotate-server-certificates="false" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354993 2588 flags.go:64] FLAG: --runonce="false" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.354999 2588 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.355005 2588 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.355010 2588 flags.go:64] FLAG: --seccomp-default="false" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.355015 2588 flags.go:64] FLAG: --serialize-image-pulls="true" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.355020 2588 flags.go:64] FLAG: --skip-headers="false" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.355025 2588 flags.go:64] FLAG: --skip-log-headers="false" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.355032 2588 flags.go:64] FLAG: --stderrthreshold="2" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.355038 2588 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.355044 2588 flags.go:64] FLAG: --storage-driver-db="cadvisor" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.355049 2588 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.355055 2588 flags.go:64] FLAG: --storage-driver-password="root" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.355061 2588 flags.go:64] FLAG: --storage-driver-secure="false" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.355066 2588 flags.go:64] FLAG: --storage-driver-table="stats" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.355071 2588 flags.go:64] FLAG: --storage-driver-user="root" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.355076 2588 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.355081 2588 flags.go:64] FLAG: --sync-frequency="1m0s" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.355086 2588 flags.go:64] FLAG: --system-cgroups="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.355091 2588 flags.go:64] FLAG: --system-reserved="cpu=500m,ephemeral-storage=1Gi,memory=1Gi" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.355101 2588 flags.go:64] FLAG: --system-reserved-cgroup="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.355106 2588 flags.go:64] FLAG: --tls-cert-file="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.355111 2588 flags.go:64] FLAG: --tls-cipher-suites="[]" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.355120 2588 flags.go:64] FLAG: --tls-min-version="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.355127 2588 flags.go:64] FLAG: --tls-private-key-file="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.355133 2588 flags.go:64] FLAG: --topology-manager-policy="none" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.355139 2588 flags.go:64] FLAG: --topology-manager-scope="container" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.355144 2588 flags.go:64] FLAG: --v="2" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.355151 2588 flags.go:64] FLAG: --version="false" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.355158 2588 flags.go:64] FLAG: --vmodule="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.355165 2588 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.355171 2588 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.355820 2588 feature_gate.go:246] feature gates: &{map[APIPriorityAndFairness:true CSIMigrationAzureFile:false CSIMigrationvSphere:false DownwardAPIHugePages:true RotateKubeletServerCertificate:true]} Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.359601 2588 server.go:413] "Kubelet version" kubeletVersion="v1.25.4+77bec7a" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.359719 2588 server.go:415] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.359782 2588 feature_gate.go:246] feature gates: &{map[APIPriorityAndFairness:true CSIMigrationAzureFile:false CSIMigrationvSphere:false DownwardAPIHugePages:true RotateKubeletServerCertificate:true]} Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.359858 2588 feature_gate.go:246] feature gates: &{map[APIPriorityAndFairness:true CSIMigrationAzureFile:false CSIMigrationvSphere:false DownwardAPIHugePages:true RotateKubeletServerCertificate:true]} Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.360510 2588 server.go:825] "Client rotation is on, will bootstrap in background" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.365123 2588 bootstrap.go:84] "Current kubeconfig file contents are still valid, no bootstrap necessary" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.365207 2588 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.365515 2588 server.go:882] "Starting client certificate rotation" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.365534 2588 certificate_manager.go:270] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.366287 2588 certificate_manager.go:270] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2023-01-06 06:46:04 +0000 UTC, rotation deadline is 2023-01-06 03:57:18.995413214 +0000 UTC Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.366310 2588 certificate_manager.go:270] kubernetes.io/kube-apiserver-client-kubelet: Waiting 18h19m38.629105558s for next certificate rotation Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.376230 2588 dynamic_cafile_content.go:119] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.376413 2588 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.378101 2588 manager.go:163] cAdvisor running in container: "/system.slice/kubelet.service" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.383687 2588 fs.go:133] Filesystem UUIDs: map[249B-6B8C:/dev/vda2 7e283ec8-4fd4-48d9-bedf-150e5fbbb374:/dev/vda3 fda260a9-581c-4391-b124-5cfcba02ef65:/dev/vda4] Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.383711 2588 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /sys/fs/cgroup:{mountpoint:/sys/fs/cgroup major:0 minor:25 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:43 fsType:tmpfs blockSize:0}] Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.383760 2588 nvidia.go:54] NVIDIA GPU metrics disabled Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.386782 2588 manager.go:212] Machine: {Timestamp:2023-01-05 09:37:40.386454932 +0000 UTC m=+0.308288398 CPUVendorID:GenuineIntel NumCores:2 NumPhysicalCores:1 NumSockets:1 CpuFrequency:2500000 MemoryCapacity:8192073728 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:2bf656612c3249b68a06a3657cb09103 SystemUUID:2bf65661-2c32-49b6-8a06-a3657cb09103 BootID:b15d5b11-7084-4ec5-8aef-b84d70720128 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:4096036864 Type:vfs Inodes:1000009 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:4096036864 Type:vfs Inodes:1000009 HasInodes:true} {Device:/sys/fs/cgroup DeviceMajor:0 DeviceMinor:25 Capacity:4096036864 Type:vfs Inodes:1000009 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:128300593152 Type:vfs Inodes:62651840 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:43 Capacity:4096036864 Type:vfs Inodes:1000009 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:128849018880 Scheduler:mq-deadline}] NetworkDevices:[{Name:br-ex MacAddress:00:16:3e:00:58:a6 Speed:0 Mtu:1500} {Name:br-int MacAddress:ba:b5:d4:e3:29:6e Speed:0 Mtu:1400} {Name:ens5 MacAddress:00:16:3e:00:58:a6 Speed:-1 Mtu:1500} {Name:genev_sys_6081 MacAddress:72:a3:61:23:bf:e8 Speed:0 Mtu:65000} {Name:ovn-k8s-mp0 MacAddress:16:90:f7:01:ce:c1 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:42:00:e8:c4:a7:2a Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:8192073728 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0 1] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0}] Caches:[{Id:0 Size:37486592 Type:Unified Level:3}]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.387007 2588 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.387258 2588 manager.go:228] Version: {KernelVersion:4.18.0-372.39.1.el8_6.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 412.86.202212170457-0 (Ootpa) DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.389612 2588 container_manager_linux.go:262] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.389723 2588 container_manager_linux.go:267] "Creating Container Manager object based on Node Config" nodeConfig={RuntimeCgroupsName:/system.slice/crio.service SystemCgroupsName:/system.slice KubeletCgroupsName: KubeletOOMScoreAdj:-999 ContainerRuntime: CgroupsPerQOS:true CgroupRoot:/ CgroupDriver:systemd KubeletRootDir:/var/lib/kubelet ProtectKernelDefaults:false NodeAllocatableConfig:{KubeReservedCgroupName: SystemReservedCgroupName: ReservedSystemCPUs: EnforceNodeAllocatable:map[pods:{}] KubeReserved:map[] SystemReserved:map[cpu:{i:{value:500 scale:-3} d:{Dec:} s:500m Format:DecimalSI} ephemeral-storage:{i:{value:1073741824 scale:0} d:{Dec:} s:1Gi Format:BinarySI} memory:{i:{value:1073741824 scale:0} d:{Dec:} s:1Gi Format:BinarySI}] HardEvictionThresholds:[{Signal:nodefs.inodesFree Operator:LessThan Value:{Quantity: Percentage:0.05} GracePeriod:0s MinReclaim:} {Signal:imagefs.available Operator:LessThan Value:{Quantity: Percentage:0.15} GracePeriod:0s MinReclaim:} {Signal:memory.available Operator:LessThan Value:{Quantity:100Mi Percentage:0} GracePeriod:0s MinReclaim:} {Signal:nodefs.available Operator:LessThan Value:{Quantity: Percentage:0.1} GracePeriod:0s MinReclaim:}]} QOSReserved:map[] ExperimentalCPUManagerPolicy:none ExperimentalCPUManagerPolicyOptions:map[] ExperimentalTopologyManagerScope:container ExperimentalCPUManagerReconcilePeriod:10s ExperimentalMemoryManagerPolicy:None ExperimentalMemoryManagerReservedMemory:[] ExperimentalPodPidsLimit:4096 EnforceCPULimits:true CPUCFSQuotaPeriod:100ms ExperimentalTopologyManagerPolicy:none} Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.389769 2588 topology_manager.go:134] "Creating topology manager with policy per scope" topologyPolicyName="none" topologyScopeName="container" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.389786 2588 container_manager_linux.go:302] "Creating device plugin manager" devicePluginEnabled=true Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.390286 2588 manager.go:127] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.390305 2588 server.go:64] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.390429 2588 state_mem.go:36] "Initialized new in-memory state store" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.390487 2588 util_unix.go:104] "Using this endpoint is deprecated, please consider using full URL format" endpoint="/var/run/crio/crio.sock" URL="unix:///var/run/crio/crio.sock" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.399598 2588 remote_runtime.go:139] "Using CRI v1 runtime API" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.399623 2588 util_unix.go:104] "Using this endpoint is deprecated, please consider using full URL format" endpoint="/var/run/crio/crio.sock" URL="unix:///var/run/crio/crio.sock" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.421214 2588 remote_image.go:95] "Using CRI v1 image API" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.421236 2588 server.go:1136] "Using root directory" path="/var/lib/kubelet" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.422690 2588 kubelet.go:393] "Attempting to sync node with API server" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.422711 2588 kubelet.go:282] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.422749 2588 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.422763 2588 kubelet.go:293] "Adding apiserver pod source" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.422777 2588 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.425344 2588 kuberuntime_manager.go:240] "Container runtime initialized" containerRuntime="cri-o" version="1.25.1-5.rhaos4.12.git6005903.el8" apiVersion="v1" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:37:40.426022 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:40.426089 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.426219 2588 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:37:40.426312 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:40.426349 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.428657 2588 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/azure-disk" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.428702 2588 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/azure-file" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.428714 2588 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/vsphere-volume" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.428748 2588 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.428760 2588 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/rbd" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.428773 2588 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/aws-ebs" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.428784 2588 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/gce-pd" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.428796 2588 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/cinder" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.429287 2588 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.429301 2588 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.429314 2588 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.429325 2588 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.429336 2588 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/secret" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.429355 2588 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.429367 2588 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/glusterfs" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.429378 2588 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/cephfs" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.429389 2588 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.429404 2588 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/fc" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.429415 2588 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.429427 2588 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/projected" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.429439 2588 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.430167 2588 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/csi" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.430306 2588 server.go:1175] "Started kubelet" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Started Kubernetes Kubelet. Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.432115 2588 server.go:155] "Starting to listen" address="0.0.0.0" port=10250 Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.435110 2588 server.go:438] "Adding debug handlers to kubelet server" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:40.436481 2588 kubelet.go:1333] "Image garbage collection failed once. Stats initialization may not have completed yet" err="failed to get imageFs info: unable to find data in memory cache" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.439059 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.439337 2588 certificate_manager.go:270] kubernetes.io/kubelet-serving: Certificate rotation is enabled Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.440137 2588 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:40.442955411Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=a6e4eabc-fc33-4f0a-a07c-aa989ae3a9a5 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.443206 2588 certificate_manager.go:270] kubernetes.io/kubelet-serving: Certificate expiration is 2023-01-06 06:46:04 +0000 UTC, rotation deadline is 2023-01-06 02:31:33.151369282 +0000 UTC Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.443226 2588 certificate_manager.go:270] kubernetes.io/kubelet-serving: Waiting 16h53m52.708145947s for next certificate rotation Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:40.446650 2588 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737601c1ba02a05", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 430260741, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 430260741, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.446806 2588 volume_manager.go:291] "The desired_state_of_world populator starts" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.446815 2588 volume_manager.go:293] "Starting Kubelet Volume Manager" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.446873 2588 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:37:40.447856 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:40.447915 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:40.447978 2588 controller.go:144] failed to ensure lease exists, will retry in 200ms, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:37:40.453316927Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=a6e4eabc-fc33-4f0a-a07c-aa989ae3a9a5 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.467609 2588 factory.go:153] Registering CRI-O factory Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.467626 2588 factory.go:55] Registering systemd factory Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.467717 2588 factory.go:103] Registering Raw factory Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.467778 2588 manager.go:1201] Started watching for new ooms in manager Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.468456 2588 manager.go:302] Starting recovery of all containers Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:40.547820 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.554207 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.598143 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.598337 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.601809 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.601983 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:40.615841 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.621961 2588 manager.go:307] Recovery completed Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:40.653350 2588 controller.go:144] failed to ensure lease exists, will retry in 400ms, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:40.654511 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.716067 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.721143 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.721171 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.721184 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.721715 2588 cpu_manager.go:214] "Starting CPU manager" policy="none" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.721831 2588 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.721928 2588 state_mem.go:36] "Initialized new in-memory state store" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.722986 2588 policy_none.go:49] "None policy: Start" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.723582 2588 memory_manager.go:168] "Starting memorymanager" policy="None" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.723604 2588 state_mem.go:35] "Initializing new in-memory state store" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.724791 2588 container_manager_linux.go:427] "Updating kernel flag" flag="vm/overcommit_memory" expectedValue=1 actualValue=0 Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.725046 2588 container_manager_linux.go:427] "Updating kernel flag" flag="kernel/panic" expectedValue=10 actualValue=0 Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.732953 2588 kubelet_network_linux.go:63] "Initialized iptables rules." protocol=IPv4 Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:40.754903 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:37:40.767290 2588 watcher.go:93] Error while processing event ("/sys/fs/cgroup/devices/kubepods.slice/kubepods-besteffort.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/devices/kubepods.slice/kubepods-besteffort.slice: no such file or directory Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.768142 2588 manager.go:273] "Starting Device Plugin manager" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.768279 2588 manager.go:447] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.768361 2588 server.go:77] "Starting device plugin registration server" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.768822 2588 plugin_watcher.go:52] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.769011 2588 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.769094 2588 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:40.769505 2588 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.792361 2588 kubelet_network_linux.go:63] "Initialized iptables rules." protocol=IPv6 Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.792383 2588 status_manager.go:161] "Starting to sync pod status with apiserver" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.792404 2588 kubelet.go:2033] "Starting kubelet main sync loop" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:40.792448 2588 kubelet.go:2057] "Skipping pod synchronization" err="PLEG is not healthy: pleg has yet to be successful" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:37:40.794156 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:40.794324 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.816875 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.817973 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.817998 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.818012 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.818034 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:40.818754 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:40.855087 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:40.893343 2588 kubelet.go:2119] "SyncLoop ADD" source="file" pods=[] Jan 05 09:37:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:40.955708 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:41.054850 2588 controller.go:144] failed to ensure lease exists, will retry in 800ms, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:41.055862 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:41.156143 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:41.219290 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:37:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:41.220474 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:37:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:41.220501 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:37:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:41.220514 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:37:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:41.220538 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:37:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:41.221315 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:37:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:41.256607 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:41.357022 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:41.441135 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:41.457539 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:41.557958 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:37:41.581826 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:41.581878 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:41.658159 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:37:41.705088 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:41.705142 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:37:41.712815 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:41.712856 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:41.759200 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:41.856074 2588 controller.go:144] failed to ensure lease exists, will retry in 1.6s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:41.860144 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:41.960625 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:42.021772 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:37:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:42.023053 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:37:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:42.023081 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:37:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:42.023095 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:37:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:42.023119 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:37:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:42.023857 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:37:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:42.061194 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:42.161597 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:42.262022 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:37:42.330982 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:42.331017 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:42.362288 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:42.441370 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:42.462826 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:42.563190 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:42.663538 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:42.763964 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:42.864292 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:42.964818 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:43.065165 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:43.165544 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:43.265966 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:43.366319 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:37:43.391416 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:43.391450 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:43.441293 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:43.457531 2588 controller.go:144] failed to ensure lease exists, will retry in 3.2s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:43.466415 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:43.566827 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:43.624937 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:37:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:43.626160 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:37:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:43.626190 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:37:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:43.626203 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:37:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:43.626254 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:37:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:43.627053 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:37:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:37:43.628689 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:43.628720 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:43.667005 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:37:43.706899 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:43.706930 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:43.767181 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:43.867470 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:43.967980 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:44.068350 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:44.168725 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:44.269133 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:44.369543 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:44.441738 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:44.470178 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:37:44.519062 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:44.519095 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:44.570374 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:44.670776 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:44.771164 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:44.871415 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:44.971940 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:45.072297 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:45.172698 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:45.273079 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:45.373530 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:45.441597 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:45.473766 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:45.574110 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:45.674483 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:45.775411 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:45.875482 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:45.975979 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:46.076350 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:46.176723 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:46.277114 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:46.377557 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:46.441553 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:46.477888 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:46.578284 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:46.659379 2588 controller.go:144] failed to ensure lease exists, will retry in 6.4s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:46.678726 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:46.779169 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:46.827105 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:37:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:46.828172 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:37:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:46.828202 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:37:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:46.828216 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:37:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:46.828251 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:37:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:46.829021 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:37:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:46.879463 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:37:46.977583 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:46.977616 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:46.979652 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:47.080042 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:47.180418 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:47.280812 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:47.381292 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:37:47.423210 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:47.423244 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:47.440977 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:47.482376 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:47.582769 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:47.683155 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:47.783654 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:47.883845 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:47.984171 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:48.084335 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:37:48.084474 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:48.084507 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:48.184772 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:48.285163 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:48.385652 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:48.441654 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:48.486042 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:48.586417 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:37:48.591304 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:48.591353 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:48.686614 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:48.787108 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:48.887415 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:48.987749 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:49.088078 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:49.188399 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:49.288782 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:49.389298 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:49.441206 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:49.489656 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:49.590036 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:49.690419 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:49.790906 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:49.891233 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:49.991611 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:50.091975 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:50.192293 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:50.253309 2588 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737601c1ba02a05", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 430260741, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 430260741, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:37:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:50.292472 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:50.393344 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:50.441275 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:50.493807 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:50.594172 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:50.694548 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:50.769784 2588 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:50.794804 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:50.894855 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:50.995178 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:51.096207 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:51.196549 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:51.296792 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:51.397275 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:51.441225 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:51.497325 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:51.598414 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:51.698779 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:51.799266 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:51.899429 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:51.999789 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:52.100152 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:52.200412 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:52.301389 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:52.401558 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:52.441474 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:52.501822 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:52.601909 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:52.702833 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:52.803288 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:52.904286 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:53.004645 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:53.060615 2588 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:53.105657 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:53.205860 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:53.229260 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:37:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:53.230519 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:37:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:53.230555 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:37:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:53.230570 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:37:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:53.230592 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:37:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:53.231396 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:37:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:53.306648 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:53.407002 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:53.440939 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:53.507242 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:53.607605 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:53.707981 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:53.808268 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:53.908737 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:54.009003 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:54.109365 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:54.209553 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:54.309980 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:54.410471 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:54.441863 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:54.511292 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:54.611718 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:54.711954 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:54.812924 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:54.913997 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:55.014166 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:55.115044 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:55.215378 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:55.315738 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:55.416106 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:55.441002 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:55.516416 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:55.616778 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:55.717156 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:55.817692 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:55.918015 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:56.018226 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:37:56.099180 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:56.099214 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:56.118299 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:56.218701 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:56.319079 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:56.419523 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:56.441530 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:56.519707 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:56.620038 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:56.720294 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:56.821349 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:56.921826 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:57.022207 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:57.122553 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:57.222955 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:37:57.225864 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:57.225899 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:57.323193 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:57.423652 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:57.441600 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:57.524026 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:57.624250 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:57.724630 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:57.824835 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:57.925292 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:58.025695 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:58.126254 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:58.226473 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:58.326703 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:58.427171 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:58.441355 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:58.527791 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:37:58.536631 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:58.536702 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:58.628516 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:58.728764 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:58.829199 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:58.929940 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:59.030297 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:37:59.112339 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:59.112388 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:59.130473 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:59.230700 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:59.330929 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:59.431414 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:37:59.441341 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:37:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:59.531822 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:59.632858 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:59.733063 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:59.833363 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:37:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:37:59.933842 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:00.034034 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:00.062021 2588 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:00.134228 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:00.232415 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:38:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:00.233643 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:38:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:00.233689 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:38:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:00.233704 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:38:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:00.233726 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:38:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:00.234397 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:00.234447 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:38:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:00.254383 2588 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737601c1ba02a05", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 430260741, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 430260741, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:38:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:00.334724 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:00.435614 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:00.441459 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:00.536682 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:00.637064 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:00.737421 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:00.770691 2588 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:00.838293 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:00.938777 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:01.039169 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:01.139558 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:01.239986 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:01.340356 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:01.440957 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:01.441792 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:01.542259 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:01.642612 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:01.743026 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:01.843352 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:01.943842 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:02.044182 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:02.144548 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:02.244955 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:02.345294 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:02.441610 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:02.445698 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:02.546219 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:02.646553 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:02.746916 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:02.847140 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:02.947467 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:03.047769 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:03.148128 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:03.248485 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:03.348831 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:03.440978 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:03.449185 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:03.549701 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:03.650031 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:03.750532 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:03.850887 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:03.951349 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:04.051769 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:04.152155 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:04.252551 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:04.352961 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:04.441033 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:04.453240 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:04.553754 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:04.654139 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:04.754546 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:04.854858 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:04.955348 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:05.055754 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:05.156150 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:05.256418 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:05.356951 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:05.441014 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:05.457341 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:05.557809 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:05.658156 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:05.758492 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:05.859141 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:05.959474 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:06.059764 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:06.160044 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:06.260239 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:06.360575 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:06.441648 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:06.461039 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:06.561409 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:06.661823 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:06.762204 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:06.863070 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:06.963537 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:07.063697 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:07.063742 2588 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:07.164180 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:07.235296 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:38:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:07.236460 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:38:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:07.236483 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:38:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:07.236497 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:38:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:07.236519 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:38:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:07.237321 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:38:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:07.264599 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:07.364838 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:07.440915 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:07.465292 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:07.565695 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:07.666066 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:07.766477 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:07.867538 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:07.968042 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:08.068406 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:08.168808 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:08.268967 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:08.369294 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:08.441364 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:08.469821 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:08.570181 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:08.670415 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:08.770823 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:08.870922 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:08.971412 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:09.071811 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:09.172155 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:09.272542 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:09.372961 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:09.440985 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:09.473348 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:09.573730 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:09.674107 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:09.774467 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:09.874819 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:09.975295 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:10.075547 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:10.175936 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:10.255984 2588 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737601c1ba02a05", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 430260741, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 430260741, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:38:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:10.276188 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:10.376687 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:10.441562 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:10.477001 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:10.577375 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:10.678072 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:10.771472 2588 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:10.778348 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:10.878593 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:10.978960 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:11.079292 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:11.179684 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:11.279908 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:11.380353 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:11.441163 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:11.480601 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:11.580974 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:11.681350 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:11.781531 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:11.882020 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:11.982277 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:12.083051 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:12.183415 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:12.283797 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:12.384142 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:12.441069 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:12.484530 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:38:12.561616 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:12.561692 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:12.584814 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:12.684861 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:12.785202 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:12.885381 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:12.985780 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:13.086150 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:13.186354 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:13.286724 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:13.387099 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:13.440965 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:13.487278 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:13.587471 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:13.687849 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:13.788212 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:13.888603 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:13.988970 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:14.065528 2588 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:14.089721 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:14.189948 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:14.238077 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:38:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:14.239343 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:38:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:14.239376 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:38:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:14.239390 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:38:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:14.239412 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:38:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:14.240186 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:38:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:14.290527 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:14.390816 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:14.441685 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:14.491180 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:14.591269 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:14.691592 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:14.792035 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:14.893017 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:14.993351 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:15.093710 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:15.194058 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:15.294407 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:15.395014 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:15.441691 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:15.495241 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:15.595590 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:15.695825 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:15.796826 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:15.897423 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:15.997776 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:16.098137 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:16.198491 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:16.298699 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:16.399162 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:16.440987 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:16.499528 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:16.599887 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:16.700244 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:16.800776 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:16.900980 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:17.001348 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:17.101706 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:17.202072 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:17.302420 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:17.402896 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:17.441754 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:17.503291 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:17.603654 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:17.704016 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:17.804350 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:17.905428 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:18.006437 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:18.106641 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:18.206882 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:18.307234 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:18.407528 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:18.441351 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:18.507735 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:18.608087 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:18.708420 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:18.808973 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:18.909232 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:19.009611 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:19.109958 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:19.210174 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:19.310532 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:19.411013 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:19.440772 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:19.511276 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:38:19.607003 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:19.607040 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:19.612072 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:19.712416 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:19.813028 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:19.913805 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:20.014178 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:20.114538 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:20.214901 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:38:20.224827 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:20.224880 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:20.256800 2588 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737601c1ba02a05", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 430260741, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 430260741, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:38:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:20.315132 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:20.415685 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:20.441609 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:20.516126 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:20.616289 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:20.716672 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:20.771794 2588 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:20.816999 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:20.918081 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:21.018415 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:21.067294 2588 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:21.119472 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:21.219713 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:21.240831 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:38:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:21.242033 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:38:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:21.242063 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:38:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:21.242076 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:38:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:21.242098 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:38:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:21.242840 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:38:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:21.320177 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:21.420409 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:21.441214 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:21.521297 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:21.621681 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:21.722044 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:21.822607 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:21.922836 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:22.023191 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:22.123550 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:22.223954 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:22.325008 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:22.425288 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:22.441139 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:22.526288 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:22.626685 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:22.727060 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:38:22.736988 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:22.737039 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:22.827577 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:22.927739 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:23.028126 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:23.128510 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:23.228783 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:23.328998 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:23.429358 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:23.441161 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:23.530347 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:23.630722 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:23.731112 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:23.831714 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:23.931900 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:24.032075 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:24.132412 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:24.232824 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:24.333220 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:24.433663 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:24.441510 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:24.534028 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:24.634289 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:24.734692 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:24.835293 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:24.935481 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:38:24.997701134Z" level=error msg="Failed to cleanup (probably retrying): failed to destroy network for pod sandbox k8s_prometheus-k8s-0_openshift-monitoring_4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879_0(04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0): error removing pod openshift-monitoring_prometheus-k8s-0 from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-monitoring/prometheus-k8s-0/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" Jan 05 09:38:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:25.036489 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:25.136899 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:25.237295 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:25.337505 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:25.438136 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:25.441009 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:38:25.498352972Z" level=info msg="cleanup sandbox network" Jan 05 09:38:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:38:25.498692002Z" level=info msg="Got pod network &{Name:prometheus-k8s-0 Namespace:openshift-monitoring ID:04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0 UID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:38:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:38:25.498850727Z" level=info msg="Deleting pod openshift-monitoring_prometheus-k8s-0 from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:38:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:25.538543 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:25.638934 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:25.739941 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:25.840821 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:25.941482 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:26.041906 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:26.142285 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:26.242709 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:26.343096 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:26.441341 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:26.443426 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:26.543912 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:26.644946 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:26.745350 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:26.845579 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:26.946067 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:27.046416 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:27.146757 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:27.246967 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:27.347101 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:27.441214 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:27.447306 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:27.547594 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:27.647891 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:27.748697 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:27.848969 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:27.949420 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:28.049822 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:28.068699 2588 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:28.150005 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:28.243292 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:38:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:28.244458 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:38:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:28.244494 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:38:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:28.244508 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:38:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:28.244531 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:38:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:28.245280 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:38:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:28.250193 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:28.350533 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:28.441579 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:28.450792 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:28.551292 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:28.651701 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:28.751950 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:28.852123 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:28.952541 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:29.052963 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:29.153242 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:29.253642 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:29.354054 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:29.441110 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:29.454407 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:29.554775 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:29.655165 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:29.755355 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:29.855537 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:29.955854 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:30.056244 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:30.156638 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:30.257649 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:30.258430 2588 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737601c1ba02a05", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 430260741, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 430260741, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:38:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:30.358031 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:30.441021 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:30.458352 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:30.558855 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:30.659253 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:30.759612 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:30.772747 2588 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:30.859708 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:30.960755 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:31.061152 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:31.161293 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:31.261482 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:31.361562 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:31.441588 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:31.462031 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:31.562399 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:31.662607 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:31.762809 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:31.863155 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:31.963400 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:32.063729 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:32.164063 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:32.264370 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:32.364413 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:32.441400 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:32.464845 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:32.565291 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:32.665615 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:32.765962 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:32.866296 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:32.966822 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:33.066979 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:33.167370 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:33.268294 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:33.368680 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:33.441442 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:33.468832 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:33.569158 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:33.669393 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:33.769819 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:33.870907 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:33.971282 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:34.072238 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:34.172615 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:34.273026 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:34.373408 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:34.441397 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:34.473725 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:34.574180 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:34.674540 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:34.774946 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:34.875275 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:34.975697 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:35.069772 2588 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:35.075787 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:35.176119 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:35.246092 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:38:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:35.247297 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:38:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:35.247324 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:38:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:35.247337 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:38:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:35.247360 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:38:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:35.248248 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:38:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:35.276414 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:35.376923 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:35.441825 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:35.477293 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:35.577696 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:35.677947 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:35.778198 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:35.879019 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:35.979393 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:36.079772 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:36.180137 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:36.280538 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:36.381026 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:36.440817 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:36.481346 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:36.581733 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:36.682118 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:36.782502 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:36.882953 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:36.983197 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:37.083538 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:37.183932 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:37.284177 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:37.384668 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:37.441373 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:37.484939 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:37.585292 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:37.685675 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:37.786048 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:37.886376 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:37.987468 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:38.087696 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:38.188061 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:38.288418 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:38.388895 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:38.441774 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:38.489353 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:38.589716 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:38.690081 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:38.790425 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:38.890834 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:38.991168 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:39.092292 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:39.192679 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:39.293043 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:39.393530 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:39.441196 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:39.493768 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:39.593940 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:39.694381 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:39.794764 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:39.895055 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:39.995412 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:40.095767 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:40.195959 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:40.259906 2588 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737601c1ba02a05", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 430260741, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 430260741, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:38:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:40.296109 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:40.396755 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:40.441710 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:40.497291 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:40.597649 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:40.698133 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:40.773077 2588 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:40.798293 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:40.898600 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:40.998962 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:41.099131 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:41.199496 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:41.299890 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:41.400345 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:41.441176 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:41.500750 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:41.601106 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:41.701407 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:41.801887 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:41.902826 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:42.002903 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:42.070920 2588 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:42.103107 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:42.203409 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:42.248603 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:38:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:42.249917 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:38:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:42.249947 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:38:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:42.249961 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:38:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:42.249985 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:38:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:42.250709 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:38:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:42.303887 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:42.404362 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:42.441186 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:42.504753 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:42.605102 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:42.705412 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:42.805900 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:42.906099 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:43.006246 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:43.106603 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:43.206947 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:43.307296 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:43.407767 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:43.441643 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:43.508281 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:43.608616 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:43.708978 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:43.809581 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:43.909781 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:44.010151 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:44.110501 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:44.210874 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:44.311200 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:44.411408 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:44.441300 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:44.511827 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:44.612141 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:44.712513 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:44.812965 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:44.913158 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:45.013514 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:45.113721 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:45.213812 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:45.314028 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:45.414409 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:45.441195 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:45.514769 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:45.615123 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:45.715466 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:45.816531 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:45.917468 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:46.017827 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:46.118178 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:46.218384 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:46.318621 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:46.419093 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:46.440895 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:46.519412 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:46.619767 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:46.720135 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:46.820531 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:46.920871 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:47.021195 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:47.121545 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:47.221910 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:47.322195 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:47.422654 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:47.441480 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:47.523208 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:47.623350 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:47.723721 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:47.824300 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:47.925177 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:48.025539 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:48.125905 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:48.226279 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:48.326652 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:48.427127 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:48.440984 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:48.527621 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:48.627987 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:48.728366 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:48.828920 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:48.929123 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:49.029517 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:49.072559 2588 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:49.129753 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:49.230129 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:49.251387 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:38:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:49.252525 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:38:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:49.252650 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:38:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:49.252666 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:38:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:49.252690 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:38:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:49.253462 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:38:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:38:49.295306 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:49.295355 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:49.330494 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:49.430971 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:49.441814 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:49.531990 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:49.632222 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:49.732540 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:49.833136 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:49.933392 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:50.033773 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:50.134152 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:50.234527 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:50.261536 2588 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737601c1ba02a05", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 430260741, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 430260741, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:38:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:50.334805 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:50.435530 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:50.441329 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:50.536543 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:50.636926 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:50.737304 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:50.773564 2588 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:38:50.792246 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:50.792306 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:50.838366 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:50.939243 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:51.039604 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:51.139779 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:51.240158 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:51.340546 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:51.441017 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:51.441726 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:51.541293 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:51.641670 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:51.742057 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:51.842643 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:51.942876 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:52.043080 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:52.143160 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:52.243359 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:52.343729 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:52.440981 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:52.444062 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:52.544276 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:52.644620 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:52.744834 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:52.845420 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:52.945614 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:53.045822 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:53.146036 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:53.246357 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:53.346697 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:53.440860 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:53.446946 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:53.547392 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:53.647678 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:53.747848 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:53.848293 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:53.948367 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:54.048443 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:54.148787 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:54.249184 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:54.349547 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:54.441580 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:54.449786 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:54.550411 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:54.650817 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:54.751239 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:54.851715 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:54.952786 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:55.053043 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:38:55.097101 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:55.097150 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:55.153416 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:55.253819 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:55.354040 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:55.441222 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:55.454130 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:55.554620 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:55.655012 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:55.755434 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:55.856521 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:55.956677 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:56.057066 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:56.073981 2588 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:56.157295 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:56.253693 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:38:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:56.254937 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:38:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:56.254964 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:38:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:56.254977 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:38:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:56.255000 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:38:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:56.255760 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:38:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:56.257841 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:56.358247 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:56.441340 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:56.458559 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:56.559024 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:56.659269 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:56.759675 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:56.860284 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:56.961293 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:57.061694 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:57.162040 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:57.262424 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:57.362685 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:57.441655 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:57.462924 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:57.563537 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:57.663933 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:57.764179 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:57.864774 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:57.965064 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:58.065414 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:58.165637 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:58.266667 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:58.366839 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:58.441803 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:58.467067 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:58.567413 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:58.667749 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:58.768034 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:58.868420 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:58.969365 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:59.069643 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:59.169708 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:59.269990 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:59.370147 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:38:59.440920 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:38:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:59.470197 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:59.570607 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:59.670948 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:59.771356 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:59.871773 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:38:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:38:59.972817 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:00.073183 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:00.173537 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:00.263309 2588 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737601c1ba02a05", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 430260741, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 430260741, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:39:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:00.274351 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:00.375022 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:00.440994 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:00.475411 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:00.575488 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:00.675875 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:00.774119 2588 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:00.776236 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:00.876749 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:00.977737 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:01.078112 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:01.178470 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:01.279009 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:01.380060 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:01.440950 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:01.480411 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:01.580905 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:01.681105 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:01.781509 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:01.881901 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:01.981988 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:02.082922 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:02.183107 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:02.283476 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:02.383976 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:02.441895 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:02.484295 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:02.584773 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:02.685151 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:02.785535 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:02.886254 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:02.987083 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:03.075231 2588 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:03.087290 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:03.187478 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:03.256740 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:39:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:03.257868 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:39:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:03.257894 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:39:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:03.257907 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:39:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:03.257930 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:39:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:03.258707 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:39:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:03.287905 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:03.388356 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:03.441159 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:03.489158 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:03.589637 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:03.690004 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:03.791036 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:03.891534 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:03.991819 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:04.092184 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:04.192419 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:04.292513 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:04.392987 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:04.441743 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:04.493192 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:04.593540 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:04.693911 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:04.794361 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:04.894747 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:04.995114 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:05.095492 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:39:05.110322 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:05.110374 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:05.195563 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:05.295938 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:05.396294 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:05.441176 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:05.496699 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:05.597063 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:05.697244 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:05.798297 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:05.898687 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:05.998980 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:06.099348 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:06.199710 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:06.300078 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:06.400529 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:06.441398 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:06.500711 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:06.600996 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:06.701360 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:06.801849 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:06.902124 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:07.002366 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:07.102724 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:07.203082 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:07.303414 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:07.403836 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:07.441538 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:07.504893 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:07.605241 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:07.705409 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:07.805712 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:07.905884 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:08.006210 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:08.106393 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:08.206585 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:08.306805 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:08.407290 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:08.441056 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:08.507619 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:08.607968 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:08.708292 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:08.808666 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:08.908994 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:09.009078 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:09.109410 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:09.209794 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:09.310164 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:09.410654 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:09.441486 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:09.511050 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:09.611404 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:09.711624 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:09.811962 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:09.912234 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:10.012292 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:10.076318 2588 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:10.112488 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:10.212849 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:10.259043 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:39:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:10.260305 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:39:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:10.260332 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:39:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:10.260347 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:39:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:10.260369 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:39:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:10.261056 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:39:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:10.264683 2588 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737601c1ba02a05", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 430260741, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 430260741, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:39:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:10.313899 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:10.414586 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:10.441343 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:39:10.510003383Z" level=error msg="Failed to cleanup (probably retrying): failed to destroy network for pod sandbox k8s_prometheus-k8s-0_openshift-monitoring_4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879_0(04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0): error removing pod openshift-monitoring_prometheus-k8s-0 from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-monitoring/prometheus-k8s-0/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" Jan 05 09:39:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:10.514945 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:10.615295 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:10.715978 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:10.774218 2588 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:10.816363 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:10.916718 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:11.017013 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:11.117370 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:11.217725 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:39:11.261101462Z" level=info msg="cleanup sandbox network" Jan 05 09:39:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:39:11.261409941Z" level=info msg="Got pod network &{Name:prometheus-k8s-0 Namespace:openshift-monitoring ID:04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0 UID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:39:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:39:11.261554725Z" level=info msg="Deleting pod openshift-monitoring_prometheus-k8s-0 from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:39:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:11.318100 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:11.418580 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:11.441457 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:11.518963 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:11.619183 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:11.719357 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:11.819846 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:11.920934 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:12.021294 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:12.121461 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:12.221828 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:12.322064 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:12.422349 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:12.441177 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:12.522616 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:12.622825 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:12.723182 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:12.823467 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:12.923535 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:13.023772 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:13.124793 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:13.225165 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:13.325548 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:13.426023 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:13.440874 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:13.526402 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:13.626763 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:13.727135 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:13.827751 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:13.928008 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:14.028232 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:14.128408 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:14.228755 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:14.329139 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:14.429633 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:14.441654 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:14.530221 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:14.630542 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:14.730919 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:14.831523 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:14.931705 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:15.032068 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:15.132407 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:15.232756 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:15.333138 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:15.433598 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:15.441376 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:15.533930 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:15.634292 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:15.734664 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:15.835586 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:15.935803 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:16.036807 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:16.137173 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:16.237538 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:16.337933 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:16.438414 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:16.441276 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:16.538773 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:16.639019 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:16.739422 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:16.840027 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:16.940195 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:17.040589 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:17.077560 2588 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:17.141605 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:17.241973 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:17.261961 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:39:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:17.263138 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:39:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:17.263167 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:39:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:17.263182 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:39:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:17.263227 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:39:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:17.263894 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:39:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:17.342969 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:17.441181 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:17.443274 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:17.543832 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:17.644217 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:17.744557 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:17.844956 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:17.945049 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:18.045406 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:18.145621 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:18.245823 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:18.346059 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:18.441132 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:18.446222 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:18.546770 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:18.647100 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:18.747351 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:18.847851 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:18.948689 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:19.049007 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:19.149239 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:19.249619 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:19.350657 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:19.441714 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:19.450923 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:19.551516 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:19.651888 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:19.752285 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:19.852771 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:19.953426 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:20.053793 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:20.154001 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:20.254250 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:20.265951 2588 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737601c1ba02a05", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 430260741, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 430260741, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:39:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:20.354421 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:20.441731 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:20.454935 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:20.555344 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:20.655738 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:20.756130 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:20.775273 2588 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:20.856715 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:20.956900 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:21.057128 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:21.157736 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:21.257979 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:21.359069 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:21.441133 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:21.459346 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:21.559794 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:21.660176 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:21.760567 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:21.861526 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:21.961629 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:22.061999 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:22.162380 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:22.262900 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:22.363293 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:22.441277 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:22.463363 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:22.563944 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:22.664290 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:22.764654 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:22.865260 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:22.965324 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:23.065714 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:23.166030 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:23.266291 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:23.366681 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:23.441450 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:23.467715 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:23.568158 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:23.668537 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:23.768932 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:23.869713 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:23.969968 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:24.070191 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:24.079141 2588 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:24.170470 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:24.264702 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:39:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:24.265917 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:39:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:24.265942 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:39:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:24.265955 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:39:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:24.265977 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:39:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:24.266765 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:39:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:24.270851 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:24.371253 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:24.441184 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:24.471505 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:24.572034 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:24.672359 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:24.772686 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:24.873164 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:24.974079 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:25.074362 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:25.174630 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:25.274909 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:25.375371 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:25.441314 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:25.475625 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:25.576051 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:25.676408 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:25.776805 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:25.877801 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:25.978864 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:26.079172 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:26.179537 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:26.279927 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:26.380241 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:26.441120 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:26.480415 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:26.581014 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:26.681393 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:26.781784 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:26.882372 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:26.983464 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:27.083833 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:27.185028 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:27.285949 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:27.386411 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:27.441243 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:27.486629 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:27.587051 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:27.687407 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:39:27.692351 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:27.692402 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:27.787668 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:27.888346 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:27.989407 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:28.089760 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:28.190120 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:28.290474 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:28.391565 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:28.441410 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:28.492630 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:28.592987 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:28.693359 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:28.793797 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:28.894369 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:28.995413 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:29.095782 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:29.196132 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:29.296498 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:29.396983 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:29.441874 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:29.497296 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:29.597569 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:29.697926 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:29.798415 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:29.899012 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:29.999052 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:30.099409 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:30.199562 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:30.267582 2588 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737601c1ba02a05", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 430260741, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 430260741, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:39:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:30.267658 2588 event.go:221] Unable to write event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737601c1ba02a05", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 430260741, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 430260741, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}' (retry limit exceeded!) Jan 05 09:39:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:30.268321 2588 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737601c25a481bd", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientMemory", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientMemory", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 598317501, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 598317501, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:39:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:30.300476 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:30.401290 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:30.441121 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:30.501536 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:30.602000 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:30.702358 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:30.775740 2588 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:30.802413 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:30.903477 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:31.004290 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:31.080376 2588 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:31.104461 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:31.204832 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:31.267075 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:39:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:31.268356 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:39:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:31.268487 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:39:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:31.268555 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:39:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:31.268619 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:39:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:31.269358 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:39:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:31.305666 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:31.406115 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:31.440968 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:31.506502 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:31.606967 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:31.707707 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:31.808192 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:31.908469 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:32.009528 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:32.109878 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:32.210115 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:32.310490 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:32.411400 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:32.441244 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:32.511683 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:32.612045 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:32.712408 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:32.812739 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:32.913250 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:33.014238 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:33.114467 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:33.214828 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:33.315246 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:33.415537 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:33.441256 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:33.516508 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:33.617176 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:33.717380 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:33.817885 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:33.918407 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:34.019436 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:34.119785 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:34.220144 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:34.320353 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:34.420845 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:34.441644 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:34.505784 2588 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737601c25a481bd", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientMemory", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientMemory", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 598317501, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 598317501, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:39:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:34.520949 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:34.621127 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:34.721492 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:34.821978 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:34.922239 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:35.023271 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:35.123942 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:35.224292 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:35.324666 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:35.425280 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:35.441064 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:35.525586 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:35.625933 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:35.726302 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:35.827176 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:35.927292 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:36.027659 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:36.127870 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:36.228171 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:36.328632 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:36.429115 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:36.440999 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:36.529415 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:36.629604 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:36.729927 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:36.830298 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:36.931124 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:37.032090 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:37.132478 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:37.232834 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:37.333210 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:37.433663 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:37.441499 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:37.534052 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:37.634240 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:37.734591 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:37.835080 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:37.935386 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:38.035708 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:38.081654 2588 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:38.135803 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:38.236166 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:38.270426 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:39:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:38.271640 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:39:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:38.271664 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:39:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:38.271677 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:39:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:38.271698 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:39:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:38.272411 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:39:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:38.336608 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:38.437095 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:38.440979 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:38.537527 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:38.637768 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:38.738136 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:38.838659 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:38.939749 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:39.040106 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:39.140496 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:39.240762 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:39.341146 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:39.441188 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:39.441245 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:39.541762 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:39.642142 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:39.742431 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:39.842484 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:39.942928 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:40.043107 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:40.143477 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:40.243691 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:40.344212 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:40.440986 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:40.445126 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:40.545596 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:40.645975 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:40.746469 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:40.775983 2588 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:40.847012 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:40.947062 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:41.047293 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:39:41.093975 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:41.094028 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:41.148252 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:41.248575 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:41.348898 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:41.440951 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:41.449160 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:41.549708 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:41.650060 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:41.750415 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:41.850855 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:41.951076 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:42.051405 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:42.151764 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:42.252155 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:42.352542 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:42.441594 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:42.452804 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:39:42.516798 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:42.516846 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:42.553054 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:42.653424 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:42.753804 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:42.854419 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:42.954615 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:43.054927 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:43.155290 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:43.255660 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:43.356036 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:43.440790 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:43.456992 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:43.557581 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:43.657948 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:43.758962 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:43.859294 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:43.959506 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:44.059819 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:44.160184 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:44.261221 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:44.361291 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:44.441382 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:44.461590 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:44.507528 2588 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737601c25a481bd", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientMemory", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientMemory", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 598317501, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 598317501, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:39:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:44.561988 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:44.662386 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:44.762769 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:44.863252 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:44.964245 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:45.064576 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:45.083421 2588 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:45.165582 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:45.265786 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:45.272903 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:39:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:45.274157 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:39:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:45.274184 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:39:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:45.274226 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:39:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:45.274248 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:39:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:45.275016 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:39:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:45.366239 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:45.441304 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:45.466562 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:45.567161 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:45.667520 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:45.767895 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:45.868834 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:45.969018 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:46.069246 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:46.169614 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:46.269980 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:46.370220 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:46.441222 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:46.470436 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:46.571047 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:46.671418 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:46.771731 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:46.872367 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:46.972465 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:47.072846 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:47.173058 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:47.273294 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:47.374295 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:47.441145 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:47.475370 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:47.575969 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:47.676366 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:47.776727 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:47.877359 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:47.978248 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:48.078627 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:48.178820 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:48.279180 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:48.379472 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:48.441426 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:48.479634 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:48.580256 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:48.680609 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:48.780850 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:48.881291 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:48.982372 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:49.082531 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:49.183598 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:49.283789 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:49.384301 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:49.441179 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:49.484477 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:49.585095 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:49.685472 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:49.785854 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:49.886363 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:49.987268 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:50.087560 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:50.187871 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:50.288136 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:50.388807 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:50.441661 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:50.489870 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:50.590274 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:50.690316 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:50.776616 2588 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:50.791246 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:50.892175 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:50.992956 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:51.093295 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:51.193486 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:51.293741 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:51.394246 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:51.441116 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:51.494410 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:51.595007 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:51.695445 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:51.795931 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:51.896490 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:51.996551 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:52.084701 2588 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:52.096772 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:52.196954 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:52.275064 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:39:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:52.276399 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:39:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:52.276515 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:39:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:52.276533 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:39:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:52.276558 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:39:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:52.277345 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:39:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:52.297500 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:52.397763 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:52.441589 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:52.497916 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:52.598881 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:52.699245 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:52.799581 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:52.900133 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:53.001175 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:53.101360 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:53.201715 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:53.302089 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:53.402299 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:53.441092 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:53.502412 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:53.602895 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:53.703257 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:53.803476 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:53.903975 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:54.004181 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:54.104532 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:54.204889 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:54.305255 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:54.405537 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:54.441361 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:54.505795 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:54.508562 2588 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737601c25a481bd", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientMemory", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientMemory", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 598317501, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 598317501, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:39:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:54.605974 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:54.706293 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:54.806590 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:54.907082 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:55.008134 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:55.108291 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:55.208636 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:55.309018 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:55.409623 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:55.441513 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:55.509958 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:55.610248 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:55.710594 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:39:55.758591 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:55.758641 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:55.811293 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:55.911840 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:56.012900 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:56.113252 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:56.213615 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:39:56.271985089Z" level=error msg="Failed to cleanup (probably retrying): failed to destroy network for pod sandbox k8s_prometheus-k8s-0_openshift-monitoring_4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879_0(04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0): error removing pod openshift-monitoring_prometheus-k8s-0 from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-monitoring/prometheus-k8s-0/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" Jan 05 09:39:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:56.313874 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:56.414294 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:56.441116 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:56.514527 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:56.615008 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:56.715362 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:56.815607 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:56.916151 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:57.017231 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:57.117525 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:57.217760 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:57.318136 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:39:57.397371298Z" level=info msg="cleanup sandbox network" Jan 05 09:39:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:39:57.397674564Z" level=info msg="Got pod network &{Name:prometheus-k8s-0 Namespace:openshift-monitoring ID:04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0 UID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:39:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:39:57.397840923Z" level=info msg="Deleting pod openshift-monitoring_prometheus-k8s-0 from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:39:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:57.418638 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:57.441479 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:57.518719 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:57.619147 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:57.719836 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:57.820141 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:57.920637 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:58.021477 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:58.121820 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:58.222171 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:58.322534 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:58.422975 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:58.441858 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:58.523294 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:58.623772 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:58.724123 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:58.824596 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:58.925219 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:59.026174 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:59.086241 2588 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:59.127139 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:59.227530 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:59.277778 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:39:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:59.278903 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:39:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:59.278930 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:39:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:59.278943 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:39:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:59.278965 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:39:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:59.279760 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:39:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:59.327960 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:59.428418 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:39:59.441352 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:39:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:59.528909 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:59.629270 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:59.729461 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:59.829952 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:39:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:39:59.930520 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:00.031568 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:00.131769 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:00.232124 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:00.332618 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:00.433472 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:00.441300 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:00.533829 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:00.634188 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:00.734570 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:00.776824 2588 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:00.835033 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:00.935599 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:01.036600 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:01.136965 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:01.237295 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:01.337667 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:01.438247 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:01.441092 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:01.538647 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:01.638876 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:01.738966 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:01.840033 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:01.940634 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:02.040717 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:02.141064 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:02.241156 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:02.341379 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:02.441427 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:02.441678 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:02.542246 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:02.643297 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:02.743667 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:02.844153 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:02.944701 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:03.045708 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:03.146726 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:03.246768 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:03.347028 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:03.441087 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:03.447252 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:03.547673 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:03.647901 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:03.748142 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:03.848523 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:03.948961 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:04.050020 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:04.150348 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:04.250566 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:04.350940 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:04.441250 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:04.451422 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:04.509444 2588 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737601c25a481bd", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientMemory", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientMemory", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 598317501, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 598317501, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:40:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:04.551777 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:04.652139 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:04.752515 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:04.853070 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:04.953527 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:05.054547 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:05.154898 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:05.255138 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:05.355354 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:05.441405 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:05.455612 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:05.555995 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:05.656358 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:05.756575 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:05.857491 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:05.957542 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:06.057808 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:06.087661 2588 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:06.157968 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:06.258009 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:06.280126 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:40:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:06.281402 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:40:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:06.281531 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:40:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:06.281548 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:40:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:06.281571 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:40:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:06.282356 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:40:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:06.358776 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:06.440817 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:06.459212 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:06.559725 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:06.660582 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:06.760938 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:06.861537 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:06.961767 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:07.062133 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:07.162492 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:07.262712 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:07.363033 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:07.441090 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:07.463296 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:07.563674 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:07.664035 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:07.764391 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:07.864790 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:07.964979 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:08.065366 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:08.165587 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:08.265955 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:08.366294 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:08.441311 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:08.466559 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:08.567022 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:08.667390 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:08.767549 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:08.868162 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:08.969125 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:09.069485 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:09.169680 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:09.270038 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:09.370429 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:09.441536 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:09.470744 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:09.571340 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:09.671712 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:09.772071 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:09.872659 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:09.972880 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:10.073215 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:10.173369 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:10.274291 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:10.374541 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:10.441643 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:10.474851 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:10.575386 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:10.676107 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:10.776489 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:10.777554 2588 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:10.876864 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:10.977949 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:11.078267 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:11.178618 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:11.278796 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:11.379048 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:11.441021 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:11.479296 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:11.579765 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:11.680089 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:11.780137 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:11.880913 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:11.980965 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:12.081137 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:12.181491 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:12.281706 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:12.382140 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:12.441132 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:12.482251 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:12.583254 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:12.683618 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:12.783980 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:12.884729 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:12.985802 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:13.086066 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:13.088979 2588 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:13.187124 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:13.283293 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:40:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:13.284536 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:40:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:13.284564 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:40:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:13.284578 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:40:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:13.284600 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:40:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:13.285340 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:40:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:13.287420 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:13.387907 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:13.441850 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:13.488054 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:13.588654 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:13.688999 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:13.789081 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:13.889722 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:13.990827 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:14.091122 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:14.191500 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:14.291745 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:14.392244 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:14.441052 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:14.492325 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:14.511116 2588 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737601c25a481bd", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientMemory", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientMemory", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 598317501, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 598317501, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:40:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:14.593372 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:14.693738 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:14.793790 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:14.894345 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:14.995389 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:15.095647 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:15.195879 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:15.296923 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:15.397496 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:15.441318 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:15.498546 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:15.599073 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:15.699245 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:15.800295 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:15.900898 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:16.001812 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:16.101981 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:16.202293 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:16.302625 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:16.403116 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:16.440980 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:16.503293 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:16.603891 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:16.704246 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:16.804736 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:16.905142 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:17.006079 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:17.106404 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:17.206761 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:17.307114 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:17.407614 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:17.441485 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:17.507811 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:17.608409 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:17.708772 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:17.809282 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:17.910299 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:18.011141 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:18.111494 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:18.211857 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:18.312169 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:18.413249 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:18.441068 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:18.513452 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:18.613928 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:18.714288 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:18.814784 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:18.915158 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:19.016242 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:19.116408 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:19.216757 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:19.317123 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:19.417469 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:19.441285 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:19.517669 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:19.617972 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:19.718292 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:19.818777 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:19.919129 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:20.020174 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:20.090296 2588 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:20.121164 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:20.221531 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:20.285742 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:40:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:20.286871 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:40:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:20.286899 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:40:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:20.286912 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:40:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:20.286935 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:40:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:20.287680 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:40:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:20.321890 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:20.422155 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:20.441015 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:20.522460 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:20.622939 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:20.723292 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:20.777621 2588 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:20.823897 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:20.924347 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:21.025006 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:21.125366 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:21.225716 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:21.326092 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:21.426477 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:21.441303 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:21.527523 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:21.627993 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:21.728345 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:21.828683 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:21 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:21.929117 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:22.030179 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:22.130530 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:22.230889 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:22.331255 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:22.431536 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:22.441345 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:22.531786 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:22.632257 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:22.732609 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:22.833096 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:22.933733 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:23.034351 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:23.134717 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:23.235071 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:23.335424 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:23.435813 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:23.441652 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:23.535878 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:23.636289 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:23.736647 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:23.837054 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:23.937296 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:24.038084 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:24.138249 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:24.238581 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:24.338944 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:24.439412 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:24.441248 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:24.512300 2588 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737601c25a481bd", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientMemory", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientMemory", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 598317501, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 598317501, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:40:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:24.539469 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:24.639826 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:24.740893 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:24.841376 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:24.941980 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:25.043033 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:25.143278 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:25.243595 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:25.344628 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:25.440953 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:25.445027 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:25.545625 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:25.645970 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:40:25.734813 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:25.734865 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:25.746886 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:25.847772 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:25.948366 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:26.049419 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:26.149782 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:26.249977 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:26.350991 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:26.440863 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:26.451085 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:26.551502 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:26.651705 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:26.752064 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:26.852353 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:26.952831 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:27.053742 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:27.091577 2588 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:27.153835 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:27.254223 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:27.288288 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:40:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:27.289520 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:40:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:27.289547 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:40:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:27.289560 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:40:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:27.289582 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:40:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:27.290980 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:40:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:27.355059 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:27.440844 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:27.456034 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:27.556533 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:27.656901 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:27.757152 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:27.857640 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:27.958270 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:28.059144 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:28.159506 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:28.259871 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:28.360251 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:28.441369 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:28.460569 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:28.561166 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:28.661532 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:28.761901 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:28.862350 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:28.962951 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:29.063721 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:29.164040 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:29.264420 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:29.364807 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:29.440896 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:29.465055 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:29.565507 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:29.665858 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:29.766218 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:29.866803 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:29.967281 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:30.068298 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:30.168663 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:30.269020 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:30.369474 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:30.441730 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:30.469937 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:30.570351 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:30.670715 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:30.771093 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:30.778220 2588 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:30.871677 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:30.972167 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:31.073236 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:40:31.163348 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:31.163399 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes?fieldSelector=metadata.name%3Dpliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl&limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:31.173410 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:31.273618 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:31.373975 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:31.441076 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:31.474292 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:31.574825 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:31.675186 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:31.775534 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:31.876135 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:31.976532 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:32.077432 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:32.177781 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:32.278142 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:32.378485 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:32.441148 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:32.479347 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:32.579880 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:40:32.605771 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:32.605824 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:32.679980 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:32.780360 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:32.881082 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:32.981347 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:33.082367 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:33.182533 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:33.282894 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:40:33.354023 2588 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:134: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:33.354072 2588 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:33.383225 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:33.441123 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:33.483306 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:33.583903 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:33.684126 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:33.784304 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:33.885050 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:33.985415 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:34.086416 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:34.093320 2588 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:34.186625 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:34.286847 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:34.291959 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:40:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:34.293146 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:40:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:34.293182 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:40:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:34.293213 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:40:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:34.293236 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:40:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:34.294036 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:40:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:34.387493 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:34.441481 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:34.488548 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:34.513382 2588 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737601c25a481bd", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientMemory", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientMemory", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 598317501, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 598317501, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:40:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:34.588706 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:34.689066 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:34.789417 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:34.890091 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:34.990490 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:35.091523 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:35.192610 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:35.292781 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:35.393292 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:35.441051 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:35.494259 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:35.594744 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:35.695044 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:35.795522 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:35.896608 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:35.996667 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:36.097015 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:36.197378 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:36.297727 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:36.398159 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:36.440993 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:36.499000 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:36.599537 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:36.699902 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:36.800351 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:36.900961 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:37.002013 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:37.102372 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:37.202595 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:37.302951 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:37.403247 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:37.441120 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:37.504297 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:37.604678 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:37.705065 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:37.805293 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:37.905856 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:38.005962 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:38.106942 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:38.207881 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:38.308216 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:38.408610 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:38.441446 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:38.508670 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:38.609066 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:38.709421 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:38.809903 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:38.910293 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:39.011291 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:39.111663 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:39.211912 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:39.311957 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:39.412611 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:39.441491 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:39.512828 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:39.613419 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:39.713793 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:39.814275 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:39.914832 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:40.015711 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:40.116034 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:40.216236 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:40.316608 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:40.417302 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:40.441494 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:40.517959 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:40.618170 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:40.718822 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:40.778449 2588 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:40.819465 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:40.920069 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:41.021071 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:41.095147 2588 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:41.121234 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:41.221386 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:41.294733 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:40:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:41.295959 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:40:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:41.295988 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:40:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:41.296003 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:40:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:41.296027 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:40:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:41.296769 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:40:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:41.321849 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:41.422296 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:41.441115 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:41.522482 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:41.623089 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:41.723476 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:41.823953 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:41.924527 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:42.025600 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:42.125916 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:42.226167 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:42.326530 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:40:42.408402420Z" level=error msg="Failed to cleanup (probably retrying): failed to destroy network for pod sandbox k8s_prometheus-k8s-0_openshift-monitoring_4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879_0(04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0): error removing pod openshift-monitoring_prometheus-k8s-0 from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-monitoring/prometheus-k8s-0/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" Jan 05 09:40:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:42.427020 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:42.440833 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:42.528061 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:42.628521 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:42.728890 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:42.829349 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:42.929961 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:43.031023 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:43.131238 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:43.231427 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:43.331797 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:43.431880 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:43.441722 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:43.532130 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:43.632559 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:43.732636 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:43.833125 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:43.933666 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:44.034755 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:40:44.096802091Z" level=info msg="cleanup sandbox network" Jan 05 09:40:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:40:44.097114896Z" level=info msg="Got pod network &{Name:prometheus-k8s-0 Namespace:openshift-monitoring ID:04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0 UID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:40:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:40:44.097313835Z" level=info msg="Deleting pod openshift-monitoring_prometheus-k8s-0 from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:40:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:44.135074 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:44.235419 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:44.335798 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:44.436474 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:44.441338 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:44.514651 2588 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl.1737601c25a481bd", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", UID:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"NodeHasSufficientMemory", Message:"Node pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl status is now: NodeHasSufficientMemory", Source:v1.EventSource{Component:"kubelet", Host:"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl"}, FirstTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 598317501, time.Local), LastTimestamp:time.Date(2023, time.January, 5, 9, 37, 40, 598317501, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/namespaces/default/events": dial tcp 10.0.99.108:6443: connect: connection refused'(may retry after sleeping) Jan 05 09:40:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:44.536860 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:44.637291 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:44.737661 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:44.838132 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:44.938613 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:45.038683 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:45.139007 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:45.239856 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:45.339954 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:45.440601 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:45.441372 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:45.540820 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:45.641102 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:45.741485 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:45.842479 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:45.943116 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:46.044164 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:46.144543 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:46.244927 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:46.345295 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:46.441381 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:46.445459 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:46.545882 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:46.646123 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:46.746423 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:46.846597 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:46.947110 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:47.048137 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:47.148452 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:47.248782 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:47.348959 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:47.441033 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:47.449244 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:47.549727 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:47.650222 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:47.750563 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:47.850881 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:47.951242 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:48.052295 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:48.096325 2588 controller.go:144] failed to ensure lease exists, will retry in 7s, error: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl?timeout=10s": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:48.152510 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:48.252890 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:48.297130 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:40:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:48.298490 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:40:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:48.298519 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:40:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:48.298532 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:40:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:48.298557 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:40:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:48.299371 2588 kubelet_node_status.go:94] "Unable to register node with API server" err="Post \"https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/api/v1/nodes\": dial tcp 10.0.99.108:6443: connect: connection refused" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:40:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:48.353296 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:48.441073 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:48.454300 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:48.554606 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:48.654856 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:48.755042 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:48.855506 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:48.956001 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:49.057055 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:49.157144 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:49.257516 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:49.357826 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:49.440922 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:49.458130 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:49.558737 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:49.659037 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:49.759421 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:49.859854 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:49.960354 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:50.061426 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:50.161731 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:50.262040 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:50.362491 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:50.441661 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:50.462868 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:50.563475 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:50.663841 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:50.764217 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:50.779296 2588 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:50.864572 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:50.965032 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:51.066072 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:51.166424 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:51.267420 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:51.367779 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:51.440825 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:51.468032 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:51.568291 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:51.668651 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:51.768998 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:51.869482 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:51.970061 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:52.071125 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:52.171922 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:52.272292 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:52.372658 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:52.441743 2588 csi_plugin.go:1032] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com:6443/apis/storage.k8s.io/v1/csinodes/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl": dial tcp 10.0.99.108:6443: connect: connection refused Jan 05 09:40:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:52.472951 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:52.573291 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:52.673504 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:52.773731 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:52.874029 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:52.974640 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:53.074709 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:53.174871 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:53.275246 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:53.375609 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:53.476369 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:53.576966 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:53.677292 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:53.777651 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:53.878211 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:53 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:53.978683 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:54.078751 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:54.179055 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:54.279246 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:54.379540 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:54.479893 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:54.580871 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:54.681232 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:54.781780 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:54.882415 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:54 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:54.983023 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:55.083926 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:55.185001 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:55.285367 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:55.299428 2588 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 09:40:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:55.300748 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 09:40:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:55.300776 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 09:40:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:55.300791 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 09:40:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:55.300824 2588 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:40:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:55.312875 2588 kubelet_node_status.go:110] "Node was previously registered" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:40:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:40:55.313050 2588 kubelet_node_status.go:75] "Successfully registered node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 09:40:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:55.385875 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:55.486523 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:55.587179 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:55.687526 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:55.787695 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:55.888599 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:55 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:55.989510 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:56.090564 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:56.191624 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:56.291844 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:56.392910 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:56.493279 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:56.594143 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:56.694505 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:56.795561 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:56.896179 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:56.997220 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:57.098217 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:57.198493 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:57.298853 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:57.399379 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:57.499737 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:57.600151 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:57.700548 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:57.801018 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:57.901414 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:58.001672 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:58.102676 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:58.203032 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:58.303259 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:58.403744 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:58.504065 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:58.604529 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:58.704902 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:58.805229 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:58 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:58.905416 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:59.005793 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:59.105884 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:59.206256 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:59.306639 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:59.407011 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:59.507234 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:59.608224 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:59.708624 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:59.808732 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:40:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:40:59.908970 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:00.009128 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:00.110086 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:00.210258 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:00.310633 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:00.411385 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:00.511881 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:00.612095 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:00.712368 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:00.779692 2588 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:00.812707 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:00 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:00.913218 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:01.013433 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:01.114434 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:01.214776 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:01.315085 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:01.415546 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:01.515923 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:01.616380 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:01.716760 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:01.817257 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:01 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:01.917885 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:02.018252 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:02.119298 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:02.219684 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:02.320056 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:02.421025 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:02.521355 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:02.621953 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:02.722292 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:02.822902 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:02.923271 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:03.023638 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:03.124704 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:03.224933 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:03.325296 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:03.425803 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:03.526023 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:03.626627 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:03.726826 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:03.827295 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:03.927908 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:04.028847 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:04.129179 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:04.229563 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:04.329942 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:04.430425 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:04.530695 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:04.631255 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:04.731645 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:04.832118 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:04.932737 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:05.032820 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:05.133161 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:05.233561 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:05.333944 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:05.434699 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:05.534781 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:05.635292 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:05.735677 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:05.836659 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:05.937255 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:06.038293 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:06.138677 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:06.238915 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:06.339294 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:06.439823 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:06.540216 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:06.640712 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:06.741099 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:06.841577 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:06.942060 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:07.042991 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:07.143357 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:07.243751 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:07.343985 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:07.444388 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:07.544875 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:07.645807 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:07.746219 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:07.846596 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:07.946954 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:08.048036 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:08.148291 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:08.248472 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:08.348787 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:08.449293 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:08.549788 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:08.650290 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:08.750690 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:08.851028 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:08.951093 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:09.052050 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:09.152422 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:09.252813 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:09.353219 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:09.453511 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:09.554001 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:09.654511 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:09.754893 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:09.855381 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:09.956019 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:10.056904 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:10.157278 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:10.257484 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:10.357768 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:10.458294 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:10.558772 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:10.659258 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:10.759945 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:10.780064 2588 eviction_manager.go:256] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:10.860250 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:10.960715 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:11.061765 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:11.161943 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:11.262292 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:11.362657 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:11.463149 2588 kubelet.go:2471] "Error getting node" err="node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" not found" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.543866 2588 apiserver.go:52] "Watching apiserver" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.559405 2588 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-multus/network-metrics-daemon-hq594 openshift-sdn/sdn-66nsp openshift-dns/dns-default-2pfzf openshift-network-diagnostics/network-check-target-xmq2g openshift-multus/multus-additional-cni-plugins-5fgxc openshift-cluster-node-tuning-operator/tuned-smwj7 openshift-dns/node-resolver-nxzr8 openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4 openshift-ingress-canary/ingress-canary-8xrbh openshift-image-registry/node-ca-xw6d2 openshift-monitoring/prometheus-k8s-0 openshift-multus/multus-4xhw7 openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn openshift-machine-config-operator/machine-config-daemon-hf8f5 openshift-monitoring/alertmanager-main-0 openshift-monitoring/telemeter-client-5598c645c4-mftdv openshift-ingress/router-default-5cf6bbdff9-tnffs openshift-monitoring/node-exporter-2r6nf openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l openshift-monitoring/kube-state-metrics-6798b548b9-xl88q openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256] Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.559478 2588 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.559563 2588 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.559636 2588 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.559750 2588 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.563417 2588 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.563896 2588 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.564158 2588 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.564747 2588 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.565186 2588 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.565687 2588 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.566248 2588 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.567100 2588 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.568349 2588 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.569236 2588 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.572613 2588 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.572722 2588 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.572808 2588 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.572916 2588 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.573088 2588 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.573316 2588 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.573546 2588 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.576660 2588 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.705353 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-6f5dcdf79b-wq256\" (UID: \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\") " pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.705522 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7www\" (UniqueName: \"kubernetes.io/projected/8f88cbe5-1221-4e6f-a6c9-f22da152b43f-kube-api-access-g7www\") pod \"node-resolver-nxzr8\" (UID: \"8f88cbe5-1221-4e6f-a6c9-f22da152b43f\") " pod="openshift-dns/node-resolver-nxzr8" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.705633 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/05c9a6ea-140a-49a2-bde5-f220b05aa252-tls-assets\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.705720 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-web-config\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.705816 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.705905 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.705990 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-system-cni-dir\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.706083 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-proxy\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-main-proxy\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.706178 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-telemeter-client\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-secret-telemeter-client\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.706317 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-adapter-prometheus-config\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-prometheus-adapter-prometheus-config\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.706426 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-lib-cni-networks-openshift-sdn\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.706505 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-default-certificate\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.706940 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-system-cni-dir\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.706977 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemeter-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-telemeter-trusted-ca-bundle\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.707010 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-registration-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.707044 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.707076 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-tls\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.707108 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-cni-bin\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.707140 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-config\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.707174 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4sjn\" (UniqueName: \"kubernetes.io/projected/659b8bb4-81b5-49bf-8127-1d326f30ba2f-kube-api-access-h4sjn\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.707229 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-config\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.707281 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.707316 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3b10ff4-a243-4a4e-b27d-f613e258666d-service-ca-bundle\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.707349 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.707393 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-openshift-state-metrics-tls\") pod \"openshift-state-metrics-6f5dcdf79b-wq256\" (UID: \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\") " pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.707426 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzfn5\" (UniqueName: \"kubernetes.io/projected/e04a47b4-aa8e-4d4e-9d45-aa8037dcb748-kube-api-access-lzfn5\") pod \"ingress-canary-8xrbh\" (UID: \"e04a47b4-aa8e-4d4e-9d45-aa8037dcb748\") " pod="openshift-ingress-canary/ingress-canary-8xrbh" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.707456 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-run-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-openshift-sdn\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.707489 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-metrics-certs\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.707519 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.707553 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6g5r\" (UniqueName: \"kubernetes.io/projected/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-kube-api-access-h6g5r\") pod \"openshift-state-metrics-6f5dcdf79b-wq256\" (UID: \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\") " pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.707584 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-cni-binary-copy\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.707615 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-wtmp\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.707643 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-etc\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.707672 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-sys\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.707706 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccjjw\" (UniqueName: \"kubernetes.io/projected/c3b10ff4-a243-4a4e-b27d-f613e258666d-kube-api-access-ccjjw\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.707738 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.707770 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-tuned-profiles-data\" (UniqueName: \"kubernetes.io/configmap/91265f01-9dcb-4046-961e-ad4c544224d7-var-lib-tuned-profiles-data\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.707806 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"non-standard-root-system-trust-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bb84a70-232c-4467-ac0b-647030155908-non-standard-root-system-trust-ca-bundle\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.707843 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxsm5\" (UniqueName: \"kubernetes.io/projected/05c9a6ea-140a-49a2-bde5-f220b05aa252-kube-api-access-kxsm5\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.707875 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98gp9\" (UniqueName: \"kubernetes.io/projected/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-kube-api-access-98gp9\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.707906 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-multus-cni-dir\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.707936 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-metrics-tls\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.707970 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-adapter-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-prometheus-adapter-audit-profiles\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.708001 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-run-ovs\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-ovs\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.708032 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/05c9a6ea-140a-49a2-bde5-f220b05aa252-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.708064 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/05c9a6ea-140a-49a2-bde5-f220b05aa252-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.708096 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.708130 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c0e47249-a5ff-460e-958d-53577387c154-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.708164 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmxtc\" (UniqueName: \"kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc\") pod \"network-check-target-xmq2g\" (UID: \"f364a949-09d7-4cfb-83ff-e532b822a557\") " pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.708216 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/a40bf20d-67a4-45e7-b701-ca5a75376c1f-audit-log\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.708250 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-tls\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.709472 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blpwr\" (UniqueName: \"kubernetes.io/projected/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-api-access-blpwr\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.709520 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cookie-secret\" (UniqueName: \"kubernetes.io/secret/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-cookie-secret\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.709552 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-host\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.709582 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8f88cbe5-1221-4e6f-a6c9-f22da152b43f-hosts-file\") pod \"node-resolver-nxzr8\" (UID: \"8f88cbe5-1221-4e6f-a6c9-f22da152b43f\") " pod="openshift-dns/node-resolver-nxzr8" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.709614 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.709644 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-etc-sysconfig\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.709682 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd-system\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-run-systemd-system\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.709714 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-device-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.709744 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-stats-auth\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.709775 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g48vj\" (UniqueName: \"kubernetes.io/projected/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-kube-api-access-g48vj\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.709810 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.709839 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/659b8bb4-81b5-49bf-8127-1d326f30ba2f-host\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.709868 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-config-volume\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.709904 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-root\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.709938 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfs2p\" (UniqueName: \"kubernetes.io/projected/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-kube-api-access-dfs2p\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.709969 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-plugin-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.710001 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-kube-etcd-client-certs\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-kube-etcd-client-certs\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.710040 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r64p\" (UniqueName: \"kubernetes.io/projected/fc45e847-8f55-4c48-8c48-4bfee2419dca-kube-api-access-4r64p\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.710071 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-mtu-migration\" (UniqueName: \"kubernetes.io/configmap/d5b0f5e6-3d70-4315-b9c5-1403036f517b-config-mtu-migration\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.710100 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-dbus\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.710134 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-state-metrics-tls\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.710165 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx2gt\" (UniqueName: \"kubernetes.io/projected/50152296-9489-4fa8-aa42-7858debe1c08-kube-api-access-zx2gt\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.710222 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-telemeter-client-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-secret-telemeter-client-kube-rbac-proxy-config\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.710254 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-web-config\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.710307 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-serving-certs-ca-bundle\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.710340 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-metrics-client-ca\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.710370 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d5b0f5e6-3d70-4315-b9c5-1403036f517b-env-overrides\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.710402 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7xz4\" (UniqueName: \"kubernetes.io/projected/91265f01-9dcb-4046-961e-ad4c544224d7-kube-api-access-b7xz4\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.710438 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.710471 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfs6g\" (UniqueName: \"kubernetes.io/projected/a40bf20d-67a4-45e7-b701-ca5a75376c1f-kube-api-access-gfs6g\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.710505 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/88745ab7-efe7-42c0-a2e8-75591518333d-tls-certificates\") pod \"prometheus-operator-admission-webhook-598468bb49-94dsn\" (UID: \"88745ab7-efe7-42c0-a2e8-75591518333d\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.710533 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sdn-metrics-certs\" (UniqueName: \"kubernetes.io/secret/d5b0f5e6-3d70-4315-b9c5-1403036f517b-sdn-metrics-certs\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.710563 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-var-run-dbus\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.710600 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05c9a6ea-140a-49a2-bde5-f220b05aa252-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.710635 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.710666 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.710699 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.710730 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-os-release\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.710761 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-kubelet-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.710795 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-oauth-cookie\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-oauth-cookie\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.710829 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c0e47249-a5ff-460e-958d-53577387c154-cni-binary-copy\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.710872 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/659b8bb4-81b5-49bf-8127-1d326f30ba2f-serviceca\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.710905 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-volume-directive-shadow\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.710935 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-os-release\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.710966 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nd4t\" (UniqueName: \"kubernetes.io/projected/d5b0f5e6-3d70-4315-b9c5-1403036f517b-kube-api-access-8nd4t\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.711003 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzzxr\" (UniqueName: \"kubernetes.io/projected/1bb84a70-232c-4467-ac0b-647030155908-kube-api-access-vzzxr\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.711041 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-querier-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c63c9be-d1ad-44b6-8b33-09af3ee314af-thanos-querier-trusted-ca-bundle\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.711074 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvz7l\" (UniqueName: \"kubernetes.io/projected/7c63c9be-d1ad-44b6-8b33-09af3ee314af-kube-api-access-pvz7l\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.711107 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.711149 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xrks\" (UniqueName: \"kubernetes.io/projected/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-kube-api-access-5xrks\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.711182 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls\" (UniqueName: \"kubernetes.io/secret/a40bf20d-67a4-45e7-b701-ca5a75376c1f-tls\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.711241 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/7c63c9be-d1ad-44b6-8b33-09af3ee314af-metrics-client-ca\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.713314 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-rootfs\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:41:11.713541 2588 watcher.go:93] Error while processing event ("/sys/fs/cgroup/devices/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91265f01_9dcb_4046_961e_ad4c544224d7.slice": 0x40000100 == IN_CREATE|IN_ISDIR): open /sys/fs/cgroup/devices/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91265f01_9dcb_4046_961e_ad4c544224d7.slice: no such file or directory Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.714092 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a40bf20d-67a4-45e7-b701-ca5a75376c1f-tmpfs\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.714137 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-cnibin\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.714188 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-run-netns\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.714242 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/05c9a6ea-140a-49a2-bde5-f220b05aa252-config-out\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.714292 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.714789 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-config-out\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.714834 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.714870 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.714911 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-serving-certs-ca-bundle\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.714947 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemeter-client-tls\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-telemeter-client-tls\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.714981 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-metrics-client-ca\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.715014 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zf2qr\" (UniqueName: \"kubernetes.io/projected/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-kube-api-access-zf2qr\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.715046 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-sys\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.715077 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5b0f5e6-3d70-4315-b9c5-1403036f517b-config\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.715107 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-modules\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-modules\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.715137 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-lib-modules\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.715170 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-grpc-tls\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.715227 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-metrics-client-ca\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.715256 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-conf\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-cni-conf\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.715341 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.715376 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-proxy-tls\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.715409 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-textfile\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.715440 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-run-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-kubernetes\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.715473 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-config-volume\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.715508 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wq4k7\" (UniqueName: \"kubernetes.io/projected/c0e47249-a5ff-460e-958d-53577387c154-kube-api-access-wq4k7\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.715541 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-cnibin\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.715574 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.715605 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-run-crio\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-crio\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.715634 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-slash\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.715671 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.721397 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.723874 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-proxy\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-proxy\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.723933 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.724176 2588 reconciler.go:169] "Reconciler: start to sync state" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:41:11.772042 2588 watcher.go:93] Error while processing event ("/sys/fs/cgroup/devices/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50152296_9489_4fa8_aa42_7858debe1c08.slice": 0x40000100 == IN_CREATE|IN_ISDIR): readdirent /sys/fs/cgroup/devices/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50152296_9489_4fa8_aa42_7858debe1c08.slice: no such file or directory Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:41:11.841967 2588 watcher.go:93] Error while processing event ("/sys/fs/cgroup/devices/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3b10ff4_a243_4a4e_b27d_f613e258666d.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/devices/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3b10ff4_a243_4a4e_b27d_f613e258666d.slice: no such file or directory Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.846800 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-cnibin\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.846845 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.846878 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-run-crio\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-crio\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.846933 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-slash\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.846971 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.847008 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.847041 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-proxy\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-proxy\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.847073 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.847130 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-6f5dcdf79b-wq256\" (UID: \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\") " pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.847163 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-g7www\" (UniqueName: \"kubernetes.io/projected/8f88cbe5-1221-4e6f-a6c9-f22da152b43f-kube-api-access-g7www\") pod \"node-resolver-nxzr8\" (UID: \"8f88cbe5-1221-4e6f-a6c9-f22da152b43f\") " pod="openshift-dns/node-resolver-nxzr8" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.847216 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/05c9a6ea-140a-49a2-bde5-f220b05aa252-tls-assets\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.847249 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-web-config\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.847338 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.847371 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.847415 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-system-cni-dir\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.847449 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-proxy\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-main-proxy\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.847479 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-telemeter-client\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-secret-telemeter-client\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.847512 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"prometheus-adapter-prometheus-config\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-prometheus-adapter-prometheus-config\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.847543 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-lib-cni-networks-openshift-sdn\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.847574 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-default-certificate\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.847608 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-system-cni-dir\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.847643 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"telemeter-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-telemeter-trusted-ca-bundle\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.847678 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-registration-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.847713 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.847744 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-tls\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.847782 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-cni-bin\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.847811 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-config\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.847843 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-h4sjn\" (UniqueName: \"kubernetes.io/projected/659b8bb4-81b5-49bf-8127-1d326f30ba2f-kube-api-access-h4sjn\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.847883 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-config\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.847919 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.847953 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3b10ff4-a243-4a4e-b27d-f613e258666d-service-ca-bundle\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.847984 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.848019 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-openshift-state-metrics-tls\") pod \"openshift-state-metrics-6f5dcdf79b-wq256\" (UID: \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\") " pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.848058 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-lzfn5\" (UniqueName: \"kubernetes.io/projected/e04a47b4-aa8e-4d4e-9d45-aa8037dcb748-kube-api-access-lzfn5\") pod \"ingress-canary-8xrbh\" (UID: \"e04a47b4-aa8e-4d4e-9d45-aa8037dcb748\") " pod="openshift-ingress-canary/ingress-canary-8xrbh" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.848092 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-run-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-openshift-sdn\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.848123 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-metrics-certs\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.848156 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.848214 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-h6g5r\" (UniqueName: \"kubernetes.io/projected/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-kube-api-access-h6g5r\") pod \"openshift-state-metrics-6f5dcdf79b-wq256\" (UID: \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\") " pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.848247 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-cni-binary-copy\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.848523 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-wtmp\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.848607 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"etc\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-etc\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.848641 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-sys\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.848689 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-ccjjw\" (UniqueName: \"kubernetes.io/projected/c3b10ff4-a243-4a4e-b27d-f613e258666d-kube-api-access-ccjjw\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.848722 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.848768 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"var-lib-tuned-profiles-data\" (UniqueName: \"kubernetes.io/configmap/91265f01-9dcb-4046-961e-ad4c544224d7-var-lib-tuned-profiles-data\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.848806 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"non-standard-root-system-trust-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bb84a70-232c-4467-ac0b-647030155908-non-standard-root-system-trust-ca-bundle\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.848853 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-kxsm5\" (UniqueName: \"kubernetes.io/projected/05c9a6ea-140a-49a2-bde5-f220b05aa252-kube-api-access-kxsm5\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.848885 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-98gp9\" (UniqueName: \"kubernetes.io/projected/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-kube-api-access-98gp9\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.848929 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-multus-cni-dir\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.848962 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-metrics-tls\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.849010 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"prometheus-adapter-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-prometheus-adapter-audit-profiles\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.849047 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-run-ovs\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-ovs\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.849091 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/05c9a6ea-140a-49a2-bde5-f220b05aa252-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.849125 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/05c9a6ea-140a-49a2-bde5-f220b05aa252-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.849169 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.849218 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c0e47249-a5ff-460e-958d-53577387c154-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.849290 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-mmxtc\" (UniqueName: \"kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc\") pod \"network-check-target-xmq2g\" (UID: \"f364a949-09d7-4cfb-83ff-e532b822a557\") " pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.849325 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/a40bf20d-67a4-45e7-b701-ca5a75376c1f-audit-log\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.849374 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-tls\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.849411 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-blpwr\" (UniqueName: \"kubernetes.io/projected/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-api-access-blpwr\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.849459 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cookie-secret\" (UniqueName: \"kubernetes.io/secret/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-cookie-secret\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.849498 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-host\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.849543 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8f88cbe5-1221-4e6f-a6c9-f22da152b43f-hosts-file\") pod \"node-resolver-nxzr8\" (UID: \"8f88cbe5-1221-4e6f-a6c9-f22da152b43f\") " pod="openshift-dns/node-resolver-nxzr8" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.849576 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.849619 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-etc-sysconfig\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.849652 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"run-systemd-system\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-run-systemd-system\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.849698 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-device-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.849733 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-stats-auth\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.849778 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-g48vj\" (UniqueName: \"kubernetes.io/projected/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-kube-api-access-g48vj\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.849827 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.849858 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/659b8bb4-81b5-49bf-8127-1d326f30ba2f-host\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.849912 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-config-volume\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.849946 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-root\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.849992 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-dfs2p\" (UniqueName: \"kubernetes.io/projected/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-kube-api-access-dfs2p\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.850027 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"plugin-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-plugin-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.850077 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-kube-etcd-client-certs\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-kube-etcd-client-certs\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.850115 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-4r64p\" (UniqueName: \"kubernetes.io/projected/fc45e847-8f55-4c48-8c48-4bfee2419dca-kube-api-access-4r64p\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.850164 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config-mtu-migration\" (UniqueName: \"kubernetes.io/configmap/d5b0f5e6-3d70-4315-b9c5-1403036f517b-config-mtu-migration\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.850223 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-dbus\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.850288 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-state-metrics-tls\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.850324 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-zx2gt\" (UniqueName: \"kubernetes.io/projected/50152296-9489-4fa8-aa42-7858debe1c08-kube-api-access-zx2gt\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.850376 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-telemeter-client-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-secret-telemeter-client-kube-rbac-proxy-config\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.850409 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-web-config\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.850458 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-serving-certs-ca-bundle\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.850493 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-metrics-client-ca\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.850537 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d5b0f5e6-3d70-4315-b9c5-1403036f517b-env-overrides\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.850573 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-b7xz4\" (UniqueName: \"kubernetes.io/projected/91265f01-9dcb-4046-961e-ad4c544224d7-kube-api-access-b7xz4\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.850623 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.850659 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-gfs6g\" (UniqueName: \"kubernetes.io/projected/a40bf20d-67a4-45e7-b701-ca5a75376c1f-kube-api-access-gfs6g\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.850708 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/88745ab7-efe7-42c0-a2e8-75591518333d-tls-certificates\") pod \"prometheus-operator-admission-webhook-598468bb49-94dsn\" (UID: \"88745ab7-efe7-42c0-a2e8-75591518333d\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.850742 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"sdn-metrics-certs\" (UniqueName: \"kubernetes.io/secret/d5b0f5e6-3d70-4315-b9c5-1403036f517b-sdn-metrics-certs\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.850788 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-var-run-dbus\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.850835 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05c9a6ea-140a-49a2-bde5-f220b05aa252-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.850874 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.850923 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.850961 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.851008 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-os-release\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.851047 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-kubelet-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.851100 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-oauth-cookie\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-oauth-cookie\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.851154 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c0e47249-a5ff-460e-958d-53577387c154-cni-binary-copy\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.851190 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/659b8bb4-81b5-49bf-8127-1d326f30ba2f-serviceca\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.851257 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-volume-directive-shadow\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.851312 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-os-release\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.851361 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-8nd4t\" (UniqueName: \"kubernetes.io/projected/d5b0f5e6-3d70-4315-b9c5-1403036f517b-kube-api-access-8nd4t\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.851402 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-vzzxr\" (UniqueName: \"kubernetes.io/projected/1bb84a70-232c-4467-ac0b-647030155908-kube-api-access-vzzxr\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.851453 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"thanos-querier-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c63c9be-d1ad-44b6-8b33-09af3ee314af-thanos-querier-trusted-ca-bundle\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.851489 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-pvz7l\" (UniqueName: \"kubernetes.io/projected/7c63c9be-d1ad-44b6-8b33-09af3ee314af-kube-api-access-pvz7l\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.851535 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.851576 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-5xrks\" (UniqueName: \"kubernetes.io/projected/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-kube-api-access-5xrks\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.851621 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tls\" (UniqueName: \"kubernetes.io/secret/a40bf20d-67a4-45e7-b701-ca5a75376c1f-tls\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.851669 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/7c63c9be-d1ad-44b6-8b33-09af3ee314af-metrics-client-ca\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.851705 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-rootfs\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.851752 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a40bf20d-67a4-45e7-b701-ca5a75376c1f-tmpfs\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.851793 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-cnibin\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.851839 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-run-netns\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.851877 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/05c9a6ea-140a-49a2-bde5-f220b05aa252-config-out\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.851925 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.851959 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-config-out\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.852006 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.852046 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.852091 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-serving-certs-ca-bundle\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.852129 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"telemeter-client-tls\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-telemeter-client-tls\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.852164 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-metrics-client-ca\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.852218 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-zf2qr\" (UniqueName: \"kubernetes.io/projected/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-kube-api-access-zf2qr\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.852252 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-sys\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.852300 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5b0f5e6-3d70-4315-b9c5-1403036f517b-config\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.852331 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-modules\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-modules\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.852363 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-lib-modules\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.852399 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-grpc-tls\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.852435 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-metrics-client-ca\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.852468 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-cni-conf\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-cni-conf\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.852506 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.852544 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-proxy-tls\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.852580 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-textfile\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.852615 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-run-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-kubernetes\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.852660 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-config-volume\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.852710 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-wq4k7\" (UniqueName: \"kubernetes.io/projected/c0e47249-a5ff-460e-958d-53577387c154-kube-api-access-wq4k7\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.853117 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-cnibin\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.853565 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.853632 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-run-crio\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-crio\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.853692 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-slash\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.860571 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.862299 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"root\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-root\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.866715 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config-mtu-migration\" (UniqueName: \"kubernetes.io/configmap/d5b0f5e6-3d70-4315-b9c5-1403036f517b-config-mtu-migration\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.866800 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-dbus\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.869192 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/05c9a6ea-140a-49a2-bde5-f220b05aa252-tls-assets\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.871013 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-serving-certs-ca-bundle\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.872259 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-metrics-client-ca\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.872503 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d5b0f5e6-3d70-4315-b9c5-1403036f517b-env-overrides\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.873321 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-var-run-dbus\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.874962 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05c9a6ea-140a-49a2-bde5-f220b05aa252-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.875741 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"plugin-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-plugin-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.876322 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-os-release\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.876404 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-kubelet-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.877497 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.879254 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-system-cni-dir\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.879395 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-volume-directive-shadow\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.879446 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-os-release\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.882016 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-lib-cni-networks-openshift-sdn\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.883231 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-system-cni-dir\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.887965 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-registration-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.888745 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-cni-bin\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.890981 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tls\" (UniqueName: \"kubernetes.io/secret/a40bf20d-67a4-45e7-b701-ca5a75376c1f-tls\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.891449 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-rootfs\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.891583 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a40bf20d-67a4-45e7-b701-ca5a75376c1f-tmpfs\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.891660 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-cnibin\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.891727 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-run-netns\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.892248 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/05c9a6ea-140a-49a2-bde5-f220b05aa252-config-out\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.897395 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-serving-certs-ca-bundle\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.903593 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-tls\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.903750 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-sys\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.904254 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5b0f5e6-3d70-4315-b9c5-1403036f517b-config\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.904438 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-modules\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-modules\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.904529 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-lib-modules\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.905457 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-config\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.906970 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-grpc-tls\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.907687 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-metrics-client-ca\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.907747 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-cni-conf\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-cni-conf\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.907987 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.908911 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3b10ff4-a243-4a4e-b27d-f613e258666d-service-ca-bundle\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.912584 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-proxy-tls\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.913220 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-textfile\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.913297 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-run-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-kubernetes\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.913654 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.914801 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-config-volume\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.914868 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-run-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-openshift-sdn\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.915603 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-config-volume\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.915793 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.916046 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.917085 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"prometheus-adapter-prometheus-config\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-prometheus-adapter-prometheus-config\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.917287 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.917452 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.917795 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-state-metrics-tls\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.918969 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.919163 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-proxy\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-proxy\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.919845 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c0e47249-a5ff-460e-958d-53577387c154-cni-binary-copy\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.919892 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/659b8bb4-81b5-49bf-8127-1d326f30ba2f-serviceca\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.920114 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"thanos-querier-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c63c9be-d1ad-44b6-8b33-09af3ee314af-thanos-querier-trusted-ca-bundle\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.924177 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-wtmp\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.924419 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"etc\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-etc\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.924501 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-sys\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.925504 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-multus-cni-dir\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.926175 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.926302 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-run-ovs\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-ovs\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.926411 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/05c9a6ea-140a-49a2-bde5-f220b05aa252-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.927069 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/05c9a6ea-140a-49a2-bde5-f220b05aa252-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.927171 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/7c63c9be-d1ad-44b6-8b33-09af3ee314af-metrics-client-ca\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.927344 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-config-out\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.927584 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"telemeter-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-telemeter-trusted-ca-bundle\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.927722 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.928428 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-metrics-client-ca\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.928816 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.929742 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c0e47249-a5ff-460e-958d-53577387c154-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.929934 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/a40bf20d-67a4-45e7-b701-ca5a75376c1f-audit-log\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.931345 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-host\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.931406 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8f88cbe5-1221-4e6f-a6c9-f22da152b43f-hosts-file\") pod \"node-resolver-nxzr8\" (UID: \"8f88cbe5-1221-4e6f-a6c9-f22da152b43f\") " pod="openshift-dns/node-resolver-nxzr8" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.932091 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-etc-sysconfig\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.932216 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"run-systemd-system\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-run-systemd-system\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.932294 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-device-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.932945 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/659b8bb4-81b5-49bf-8127-1d326f30ba2f-host\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:41:11.933132 2588 watcher.go:93] Error while processing event ("/sys/fs/cgroup/devices/kubepods.slice/kubepods-burstable.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/devices/kubepods.slice/kubepods-burstable.slice: no such file or directory Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.939348 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.945121 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-cni-binary-copy\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.946568 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-openshift-state-metrics-tls\") pod \"openshift-state-metrics-6f5dcdf79b-wq256\" (UID: \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\") " pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.947033 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"telemeter-client-tls\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-telemeter-client-tls\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.947525 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-default-certificate\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.947927 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-proxy\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-main-proxy\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.948139 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-web-config\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.948632 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-metrics-certs\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.949257 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"var-lib-tuned-profiles-data\" (UniqueName: \"kubernetes.io/configmap/91265f01-9dcb-4046-961e-ad4c544224d7-var-lib-tuned-profiles-data\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.949953 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"non-standard-root-system-trust-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bb84a70-232c-4467-ac0b-647030155908-non-standard-root-system-trust-ca-bundle\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.951113 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"sdn-metrics-certs\" (UniqueName: \"kubernetes.io/secret/d5b0f5e6-3d70-4315-b9c5-1403036f517b-sdn-metrics-certs\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.955950 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.956481 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.963287 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/88745ab7-efe7-42c0-a2e8-75591518333d-tls-certificates\") pod \"prometheus-operator-admission-webhook-598468bb49-94dsn\" (UID: \"88745ab7-efe7-42c0-a2e8-75591518333d\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.963743 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-oauth-cookie\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-oauth-cookie\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.966849 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-kube-etcd-client-certs\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-kube-etcd-client-certs\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.967109 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-6f5dcdf79b-wq256\" (UID: \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\") " pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.967665 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-telemeter-client-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-secret-telemeter-client-kube-rbac-proxy-config\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.968103 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-wq4k7\" (UniqueName: \"kubernetes.io/projected/c0e47249-a5ff-460e-958d-53577387c154-kube-api-access-wq4k7\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.970102 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"prometheus-adapter-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-prometheus-adapter-audit-profiles\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.975159 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-web-config\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.975574 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.979322 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.983932 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-config\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.986480 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-metrics-tls\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.986749 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-telemeter-client\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-secret-telemeter-client\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.987418 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:11.994557 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.003796 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-stats-auth\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.004231 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.004521 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.004786 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-tls\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.005190 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.009469 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfs2p\" (UniqueName: \"kubernetes.io/projected/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-kube-api-access-dfs2p\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.010989 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cookie-secret\" (UniqueName: \"kubernetes.io/secret/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-cookie-secret\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.014234 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r64p\" (UniqueName: \"kubernetes.io/projected/fc45e847-8f55-4c48-8c48-4bfee2419dca-kube-api-access-4r64p\") pod \"telemeter-client-5598c645c4-mftdv\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.014607 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7www\" (UniqueName: \"kubernetes.io/projected/8f88cbe5-1221-4e6f-a6c9-f22da152b43f-kube-api-access-g7www\") pod \"node-resolver-nxzr8\" (UID: \"8f88cbe5-1221-4e6f-a6c9-f22da152b43f\") " pod="openshift-dns/node-resolver-nxzr8" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.016876 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx2gt\" (UniqueName: \"kubernetes.io/projected/50152296-9489-4fa8-aa42-7858debe1c08-kube-api-access-zx2gt\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.022329 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7xz4\" (UniqueName: \"kubernetes.io/projected/91265f01-9dcb-4046-961e-ad4c544224d7-kube-api-access-b7xz4\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.023611 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfs6g\" (UniqueName: \"kubernetes.io/projected/a40bf20d-67a4-45e7-b701-ca5a75376c1f-kube-api-access-gfs6g\") pod \"prometheus-adapter-5894cc86c7-qhmbh\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.028246 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.029045725Z" level=info msg="Running pod sandbox: openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn/POD" id=54140d90-7eb5-4850-b607-962c85c502b6 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.029097033Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.041996 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nd4t\" (UniqueName: \"kubernetes.io/projected/d5b0f5e6-3d70-4315-b9c5-1403036f517b-kube-api-access-8nd4t\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.072951 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.075405025Z" level=info msg="Running pod sandbox: openshift-multus/network-metrics-daemon-hq594/POD" id=b353c9f6-2798-4d08-8abf-2da0f599a503 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.075572119Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.078388 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzzxr\" (UniqueName: \"kubernetes.io/projected/1bb84a70-232c-4467-ac0b-647030155908-kube-api-access-vzzxr\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.078406682Z" level=info msg="Got pod network &{Name:prometheus-operator-admission-webhook-598468bb49-94dsn Namespace:openshift-monitoring ID:a4558df3b0fcc8b008b37af068d5302c4bced1a959f7e618f9a45abaabe07264 UID:88745ab7-efe7-42c0-a2e8-75591518333d NetNS:/var/run/netns/90cf1dbd-7872-4b0c-9311-93ffbf2b4bf0 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.078438440Z" level=info msg="Adding pod openshift-monitoring_prometheus-operator-admission-webhook-598468bb49-94dsn to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:41:12.089466 2588 watcher.go:93] Error while processing event ("/sys/fs/cgroup/devices/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ecd2e2c_ab52_42e0_8fde_64e45804cc9e.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/devices/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ecd2e2c_ab52_42e0_8fde_64e45804cc9e.slice: no such file or directory Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.093657 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvz7l\" (UniqueName: \"kubernetes.io/projected/7c63c9be-d1ad-44b6-8b33-09af3ee314af-kube-api-access-pvz7l\") pod \"thanos-querier-6f4f5fb7cf-rdm9l\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.096503158Z" level=info msg="Got pod network &{Name:network-metrics-daemon-hq594 Namespace:openshift-multus ID:9994a437ea1715d6af20f45f2447fa709e3ded31fbff0b6e4cbdabe018ade0c1 UID:50152296-9489-4fa8-aa42-7858debe1c08 NetNS:/var/run/netns/c543a3a8-aa1b-4690-abc6-2065756d32be Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.096529292Z" level=info msg="Adding pod openshift-multus_network-metrics-daemon-hq594 to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.102665 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xrks\" (UniqueName: \"kubernetes.io/projected/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-kube-api-access-5xrks\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.102885 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.103415434Z" level=info msg="Running pod sandbox: openshift-sdn/sdn-66nsp/POD" id=4d7070ae-0ff7-4518-8254-904cc2903d5b name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.103544876Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.116972 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.117370163Z" level=info msg="Running pod sandbox: openshift-monitoring/telemeter-client-5598c645c4-mftdv/POD" id=d77993bd-a749-4e24-b85d-15b672ca657a name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.117407854Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.123125 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-zf2qr\" (UniqueName: \"kubernetes.io/projected/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-kube-api-access-zf2qr\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.133587100Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=4d7070ae-0ff7-4518-8254-904cc2903d5b name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.141987 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4sjn\" (UniqueName: \"kubernetes.io/projected/659b8bb4-81b5-49bf-8127-1d326f30ba2f-kube-api-access-h4sjn\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.146350527Z" level=info msg="Got pod network &{Name:telemeter-client-5598c645c4-mftdv Namespace:openshift-monitoring ID:e386c433c8b6445e78f0a12c96e3583a65dc80ffd33df16f4209a33e828520f0 UID:fc45e847-8f55-4c48-8c48-4bfee2419dca NetNS:/var/run/netns/aed17900-4e8c-44c2-8afd-9f1dab49d0cf Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.146383922Z" level=info msg="Adding pod openshift-monitoring_telemeter-client-5598c645c4-mftdv to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.153359257Z" level=info msg="Ran pod sandbox 06a6d838dd853ce187ccf364676fd4e36976cae02f20a98aeda4b59df6df2d93 with infra container: openshift-sdn/sdn-66nsp/POD" id=4d7070ae-0ff7-4518-8254-904cc2903d5b name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.157735950Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:30387b6f2e605bb8d018e0c09f9d0275048fc5e7966137fbe6bd4307c88d3fdd" id=c61161de-9941-4a4b-b98b-1862db146298 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.158075826Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:cff39d57d2c205a584f8bf4299b1e125fa1a96ac92f680117ba82c20334754d1,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:30387b6f2e605bb8d018e0c09f9d0275048fc5e7966137fbe6bd4307c88d3fdd],Size_:701241664,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=c61161de-9941-4a4b-b98b-1862db146298 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:12.158755 2588 kuberuntime_manager.go:862] container &Container{Name:sdn,Image:registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:30387b6f2e605bb8d018e0c09f9d0275048fc5e7966137fbe6bd4307c88d3fdd,Command:[/bin/bash -c #!/bin/bash Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: set -euo pipefail Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # if another process is listening on the cni-server socket, wait until it exits Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: trap 'kill $(jobs -p); rm -f /etc/cni/net.d/80-openshift-network.conf ; exit 0' TERM Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: retries=0 Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: while true; do Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: if echo 'test' | socat - UNIX-CONNECT:/var/run/openshift-sdn/cniserver/socket &>/dev/null; then Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: echo "warning: Another process is currently listening on the CNI socket, waiting 15s ..." 2>&1 Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: sleep 15 & wait Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: (( retries += 1 )) Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: else Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: break Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: fi Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: if [[ "${retries}" -gt 40 ]]; then Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: echo "error: Another process is currently listening on the CNI socket, exiting" 2>&1 Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: exit 1 Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: fi Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: done Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # local environment overrides Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: if [[ -f /etc/sysconfig/openshift-sdn ]]; then Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: set -o allexport Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: source /etc/sysconfig/openshift-sdn Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: set +o allexport Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: fi Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: #BUG: cdc accidentally mounted /etc/sysconfig/openshift-sdn as DirectoryOrCreate; clean it up so we can ultimately mount /etc/sysconfig/openshift-sdn as FileOrCreate Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # Once this is released, then we can mount it properly Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: if [[ -d /etc/sysconfig/openshift-sdn ]]; then Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: rmdir /etc/sysconfig/openshift-sdn || true Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: fi Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # configmap-based overrides Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: if [[ -f /env/${K8S_NODE_NAME} ]]; then Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: set -o allexport Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: source /env/${K8S_NODE_NAME} Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: set +o allexport Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: fi Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # Take over network functions on the node Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: rm -f /etc/cni/net.d/80-openshift-network.conf Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: cp -f /opt/cni/bin/openshift-sdn /host-cni-bin/ Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: mtu_override_flag= Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: if [[ -f /config-mtu-migration/mtu.yaml ]]; then Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: mtu_override_flag="--mtu-override /config-mtu-migration/mtu.yaml" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: fi Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # Launch the network process Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: exec /usr/bin/openshift-sdn-node \ Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: --node-name ${K8S_NODE_NAME} --node-ip ${K8S_NODE_IP} \ Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: --platform-type AlibabaCloud \ Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: --proxy-config /config/kube-proxy-config.yaml \ Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: ${mtu_override_flag} \ Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: --v ${OPENSHIFT_SDN_LOG_LEVEL:-2} Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:healthz,HostPort:10256,ContainerPort:10256,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:6443,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com,ValueFrom:nil,},EnvVar{Name:OPENSHIFT_DNS_DOMAIN,Value:cluster.local,ValueFrom:nil,},EnvVar{Name:K8S_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:K8S_NODE_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.hostIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{209715200 0} {} BinarySI},},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:config-mtu-migration,ReadOnly:true,MountPath:/config-mtu-migration,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:host-var-run-crio,ReadOnly:false,MountPath:/var/run/crio,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:host-var-run-dbus,ReadOnly:true,MountPath:/var/run/dbus/,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:host-var-run-ovs,ReadOnly:true,MountPath:/var/run/openvswitch/,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:host-var-run-kubernetes,ReadOnly:true,MountPath:/var/run/kubernetes/,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:host-run-netns,ReadOnly:true,MountPath:/run/netns,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,},VolumeMount{Name:host-var-run-openshift-sdn,ReadOnly:false,MountPath:/var/run/openshift-sdn,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,},VolumeMount{Name:host-cni-bin,ReadOnly:false,MountPath:/host-cni-bin,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:host-cni-conf,ReadOnly:false,MountPath:/etc/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:host-var-lib-cni-networks-openshift-sdn,ReadOnly:false,MountPath:/var/lib/cni/networks/openshift-sdn,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:host-modules,ReadOnly:true,MountPath:/lib/modules,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:etc-sysconfig,ReadOnly:true,MountPath:/etc/sysconfig,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:kube-api-access-8nd4t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[test -f /etc/cni/net.d/80-openshift-network.conf],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[rm -f /etc/cni/net.d/80-openshift-network.conf /host-cni-bin/openshift-sdn],},HTTPGet:nil,TCPSocket:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,} start failed in pod sdn-66nsp_openshift-sdn(d5b0f5e6-3d70-4315-b9c5-1403036f517b): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.160415876Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=7d5670f1-cdb8-4a9a-8ca2-76cbed134364 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.160717963Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=7d5670f1-cdb8-4a9a-8ca2-76cbed134364 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:12.162222 2588 kuberuntime_manager.go:862] container &Container{Name:kube-rbac-proxy,Image:registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9,Command:[/bin/bash -c #!/bin/bash Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: set -euo pipefail Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: TLS_PK=/etc/pki/tls/metrics-certs/tls.key Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: TLS_CERT=/etc/pki/tls/metrics-certs/tls.crt Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # As the secret mount is optional we must wait for the files to be present. Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # The service is created in monitor.yaml and this is created in sdn.yaml. Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # If it isn't created there is probably an issue so we want to crashloop. Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: TS=$(date +%s) Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: WARN_TS=$(( ${TS} + $(( 20 * 60)) )) Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: HAS_LOGGED_INFO=0 Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: log_missing_certs(){ Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: CUR_TS=$(date +%s) Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: if [[ "${CUR_TS}" -gt "WARN_TS" ]]; then Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: echo $(date -Iseconds) WARN: sdn-metrics-certs not mounted after 20 minutes. Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: elif [[ "${HAS_LOGGED_INFO}" -eq 0 ]] ; then Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: echo $(date -Iseconds) INFO: sdn-metrics-certs not mounted. Waiting 20 minutes. Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: HAS_LOGGED_INFO=1 Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: fi Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: } Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: while [[ ! -f "${TLS_PK}" || ! -f "${TLS_CERT}" ]] ; do Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: log_missing_certs Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: sleep 5 Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: done Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: echo $(date -Iseconds) INFO: sdn-metrics-certs mounted, starting kube-rbac-proxy Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: exec /usr/bin/kube-rbac-proxy \ Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: --logtostderr \ Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: --secure-listen-address=:9101 \ Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 \ Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: --upstream=http://127.0.0.1:29101/ \ Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: --tls-private-key-file=${TLS_PK} \ Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: --tls-cert-file=${TLS_CERT} Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:9101,ContainerPort:9101,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{20971520 0} {} 20Mi BinarySI},},},VolumeMounts:[]VolumeMount{VolumeMount{Name:sdn-metrics-certs,ReadOnly:true,MountPath:/etc/pki/tls/metrics-certs,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:kube-api-access-8nd4t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,} start failed in pod sdn-66nsp_openshift-sdn(d5b0f5e6-3d70-4315-b9c5-1403036f517b): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:12.162334 2588 pod_workers.go:965] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"sdn\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-sdn/sdn-66nsp" podUID=d5b0f5e6-3d70-4315-b9c5-1403036f517b Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.167510 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzfn5\" (UniqueName: \"kubernetes.io/projected/e04a47b4-aa8e-4d4e-9d45-aa8037dcb748-kube-api-access-lzfn5\") pod \"ingress-canary-8xrbh\" (UID: \"e04a47b4-aa8e-4d4e-9d45-aa8037dcb748\") " pod="openshift-ingress-canary/ingress-canary-8xrbh" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.181703 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6g5r\" (UniqueName: \"kubernetes.io/projected/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-kube-api-access-h6g5r\") pod \"openshift-state-metrics-6f5dcdf79b-wq256\" (UID: \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\") " pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.200838 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccjjw\" (UniqueName: \"kubernetes.io/projected/c3b10ff4-a243-4a4e-b27d-f613e258666d-kube-api-access-ccjjw\") pod \"router-default-5cf6bbdff9-tnffs\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.209552 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.209889505Z" level=info msg="Running pod sandbox: openshift-monitoring/node-exporter-2r6nf/POD" id=cc0d7a34-1399-4434-9141-581dd137449e name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.209941509Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.219244411Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=cc0d7a34-1399-4434-9141-581dd137449e name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.223763 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxsm5\" (UniqueName: \"kubernetes.io/projected/05c9a6ea-140a-49a2-bde5-f220b05aa252-kube-api-access-kxsm5\") pod \"alertmanager-main-0\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:41:12.226425 2588 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa84cbad_3d0f_43bd_a8a9_46dc1cd9c57f.slice/crio-acd08c77f0076092e020bedcc20e2c94faab21df9b7d5516eba6f0dbe370935b.scope WatchSource:0}: Error finding container acd08c77f0076092e020bedcc20e2c94faab21df9b7d5516eba6f0dbe370935b: Status 404 returned error can't find the container with id acd08c77f0076092e020bedcc20e2c94faab21df9b7d5516eba6f0dbe370935b Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.227476698Z" level=info msg="Ran pod sandbox acd08c77f0076092e020bedcc20e2c94faab21df9b7d5516eba6f0dbe370935b with infra container: openshift-monitoring/node-exporter-2r6nf/POD" id=cc0d7a34-1399-4434-9141-581dd137449e name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.228090183Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180" id=4f8fa104-3a25-4c8d-8c10-26d1192e95f7 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.228339657Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b4cde205857c1e0a984b38c61275a8370136432c7c84dbfbd7f4835f26bbb935,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180],Size_:297268865,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=4f8fa104-3a25-4c8d-8c10-26d1192e95f7 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:12.228601 2588 kuberuntime_manager.go:862] init container &Container{Name:init-textfile,Image:registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180,Command:[/bin/sh -c [[ ! -d /node_exporter/collectors/init ]] || find /node_exporter/collectors/init -perm /111 -type f -exec {} \;],Args:[],WorkingDir:/var/node_exporter/textfile,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:TMPDIR,Value:/tmp,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{1 -3} {} 1m DecimalSI},memory: {{1048576 0} {} 1Mi BinarySI},},},VolumeMounts:[]VolumeMount{VolumeMount{Name:node-exporter-textfile,ReadOnly:false,MountPath:/var/node_exporter/textfile,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:node-exporter-wtmp,ReadOnly:true,MountPath:/var/log/wtmp,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:kube-api-access-dfs2p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,} start failed in pod node-exporter-2r6nf_openshift-monitoring(aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:12.228641 2588 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init-textfile\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-monitoring/node-exporter-2r6nf" podUID=aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.241236 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.241910 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-98gp9\" (UniqueName: \"kubernetes.io/projected/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-kube-api-access-98gp9\") pod \"prometheus-k8s-0\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.242154930Z" level=info msg="Running pod sandbox: openshift-monitoring/alertmanager-main-0/POD" id=0442c4c3-31eb-4586-8d53-4bf70c6ead84 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.242224975Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.247595 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.247868185Z" level=info msg="Running pod sandbox: openshift-multus/multus-additional-cni-plugins-5fgxc/POD" id=a4c3410f-983c-4bae-849f-330a41b9a4f6 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.247986951Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.253386472Z" level=info msg="Got pod network &{Name:alertmanager-main-0 Namespace:openshift-monitoring ID:f2fab4d64b6bfb23dbb4eccdef95ed2b1090988efca6d9d87c9936f1c4416a63 UID:05c9a6ea-140a-49a2-bde5-f220b05aa252 NetNS:/var/run/netns/171a3b30-0a69-40cf-bf44-3484be0306ea Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.253425201Z" level=info msg="Adding pod openshift-monitoring_alertmanager-main-0 to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.263537715Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=a4c3410f-983c-4bae-849f-330a41b9a4f6 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.268664 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.269497946Z" level=info msg="Running pod sandbox: openshift-machine-config-operator/machine-config-daemon-hf8f5/POD" id=0a9cdfd2-271d-4635-9753-f2a7c7ec4fe2 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.269536131Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.273158 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-4xhw7" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.275292956Z" level=info msg="Running pod sandbox: openshift-multus/multus-4xhw7/POD" id=c727b60f-fa77-4c38-919e-da915a8e533a name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.275415442Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:41:12.275723 2588 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0e47249_a5ff_460e_958d_53577387c154.slice/crio-e9053289b0b05a28538d40135f5851bcbb4fef2e500409bd4a7c787c89dc9515.scope WatchSource:0}: Error finding container e9053289b0b05a28538d40135f5851bcbb4fef2e500409bd4a7c787c89dc9515: Status 404 returned error can't find the container with id e9053289b0b05a28538d40135f5851bcbb4fef2e500409bd4a7c787c89dc9515 Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.277017365Z" level=info msg="Ran pod sandbox e9053289b0b05a28538d40135f5851bcbb4fef2e500409bd4a7c787c89dc9515 with infra container: openshift-multus/multus-additional-cni-plugins-5fgxc/POD" id=a4c3410f-983c-4bae-849f-330a41b9a4f6 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.278160600Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:3fc008580e4e404960b2de4020a2939ab6421bdf8f51c0b60af4ad903272c741" id=8878a94d-f688-4b36-a88c-a4b27b6cd5c6 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.278496745Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:6bb2a44a1f0506c18ef471cac7f7e8991d1e6b9d649ccbce50f751dd2863d06d,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:3fc008580e4e404960b2de4020a2939ab6421bdf8f51c0b60af4ad903272c741],Size_:382411122,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=8878a94d-f688-4b36-a88c-a4b27b6cd5c6 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:12.278946 2588 kuberuntime_manager.go:862] init container &Container{Name:egress-router-binary-copy,Image:registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:3fc008580e4e404960b2de4020a2939ab6421bdf8f51c0b60af4ad903272c741,Command:[/entrypoint/cnibincopy.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL7_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel7/bin/,ValueFrom:nil,},EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/bin/,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:os-release,ReadOnly:true,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:kube-api-access-wq4k7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,} start failed in pod multus-additional-cni-plugins-5fgxc_openshift-multus(c0e47249-a5ff-460e-958d-53577387c154): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:12.278991 2588 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"egress-router-binary-copy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" podUID=c0e47249-a5ff-460e-958d-53577387c154 Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.282324009Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=0a9cdfd2-271d-4635-9753-f2a7c7ec4fe2 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.287541 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmxtc\" (UniqueName: \"kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc\") pod \"network-check-target-xmq2g\" (UID: \"f364a949-09d7-4cfb-83ff-e532b822a557\") " pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:41:12.292919 2588 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod997b7ab4_939e_465c_9c7d_4a2ebe3a797a.slice/crio-ae18be6d794af60a14ce35ff3392698b0b7e426a5128047f31455ade1f2b79dd.scope WatchSource:0}: Error finding container ae18be6d794af60a14ce35ff3392698b0b7e426a5128047f31455ade1f2b79dd: Status 404 returned error can't find the container with id ae18be6d794af60a14ce35ff3392698b0b7e426a5128047f31455ade1f2b79dd Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.293699393Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=c727b60f-fa77-4c38-919e-da915a8e533a name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.294061 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.294677442Z" level=info msg="Running pod sandbox: openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh/POD" id=881e2e19-4071-4b0d-9e73-71cd5edabda9 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.294818214Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.296406170Z" level=info msg="Ran pod sandbox ae18be6d794af60a14ce35ff3392698b0b7e426a5128047f31455ade1f2b79dd with infra container: openshift-machine-config-operator/machine-config-daemon-hf8f5/POD" id=0a9cdfd2-271d-4635-9753-f2a7c7ec4fe2 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.297047749Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cb9edf2251410c340a6125ff331924ab57a72ad3166bd9e1305b6d3030284990" id=b4c9b9a1-26e5-42f7-b925-b3fa4245982b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.297256301Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2b208acf1f002c143fdb248dbee2517f7b6d168def4cfef1f3188939123d0dc3,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cb9edf2251410c340a6125ff331924ab57a72ad3166bd9e1305b6d3030284990],Size_:536350945,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b4c9b9a1-26e5-42f7-b925-b3fa4245982b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:12.297592 2588 kuberuntime_manager.go:862] container &Container{Name:machine-config-daemon,Image:registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cb9edf2251410c340a6125ff331924ab57a72ad3166bd9e1305b6d3030284990,Command:[/usr/bin/machine-config-daemon],Args:[start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},},VolumeMounts:[]VolumeMount{VolumeMount{Name:rootfs,ReadOnly:false,MountPath:/rootfs,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:kube-api-access-5xrks,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,} start failed in pod machine-config-daemon-hf8f5_openshift-machine-config-operator(997b7ab4-939e-465c-9c7d-4a2ebe3a797a): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.297970812Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=cabe59f6-bcf2-45c5-add7-ce831316edc2 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.298146630Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=cabe59f6-bcf2-45c5-add7-ce831316edc2 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:12.298524 2588 kuberuntime_manager.go:862] container &Container{Name:oauth-proxy,Image:registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b,Command:[],Args:[--https-address=:9001 --provider=openshift --openshift-service-account=machine-config-daemon --upstream=http://127.0.0.1:8797 --tls-cert=/etc/tls/private/tls.crt --tls-key=/etc/tls/private/tls.key --cookie-secret-file=/etc/tls/cookie-secret/cookie-secret --openshift-sar={"resource": "namespaces", "verb": "get"} --openshift-delegate-urls={"/": {"resource": "namespaces", "verb": "get"}}],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:9001,ContainerPort:9001,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},},VolumeMounts:[]VolumeMount{VolumeMount{Name:proxy-tls,ReadOnly:false,MountPath:/etc/tls/private,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:cookie-secret,ReadOnly:false,MountPath:/etc/tls/cookie-secret,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:kube-api-access-5xrks,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,} start failed in pod machine-config-daemon-hf8f5_openshift-machine-config-operator(997b7ab4-939e-465c-9c7d-4a2ebe3a797a): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:12.298921 2588 pod_workers.go:965] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"machine-config-daemon\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"oauth-proxy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" podUID=997b7ab4-939e-465c-9c7d-4a2ebe3a797a Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.302448 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-blpwr\" (UniqueName: \"kubernetes.io/projected/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-api-access-blpwr\") pod \"kube-state-metrics-6798b548b9-xl88q\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:41:12.302691 2588 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod84f9cd8a_5e58_4aff_aedd_906ed37dc97d.slice/crio-bb5b6562ae64906e3450c53a90bf4df6ea55af8ce39868944ba2cdcf7eab1188.scope WatchSource:0}: Error finding container bb5b6562ae64906e3450c53a90bf4df6ea55af8ce39868944ba2cdcf7eab1188: Status 404 returned error can't find the container with id bb5b6562ae64906e3450c53a90bf4df6ea55af8ce39868944ba2cdcf7eab1188 Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.307765855Z" level=info msg="Ran pod sandbox bb5b6562ae64906e3450c53a90bf4df6ea55af8ce39868944ba2cdcf7eab1188 with infra container: openshift-multus/multus-4xhw7/POD" id=c727b60f-fa77-4c38-919e-da915a8e533a name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.308536378Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=74150550-708d-4c97-9c1e-7625ae78590f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.308767112Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=74150550-708d-4c97-9c1e-7625ae78590f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:12.309139 2588 kuberuntime_manager.go:862] container &Container{Name:kube-multus,Image:registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f,Command:[/bin/bash -ec --],Args:[/entrypoint/cnibincopy.sh; exec /entrypoint.sh --multus-conf-file=auto --multus-autoconfig-dir=/host/var/run/multus/cni/net.d --multus-kubeconfig-file-host=/etc/kubernetes/cni/net.d/multus.d/multus.kubeconfig --readiness-indicator-file=/var/run/multus/cni/net.d/80-openshift-network.conf --cleanup-config-on-exit=true --namespace-isolation=true --multus-log-level=verbose --cni-version=0.3.1 --additional-bin-dir=/opt/multus/bin --skip-multus-binary-copy=true - "--global-namespaces=default,openshift-multus,openshift-sriov-network-operator" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: ],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL7_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel7/bin/,ValueFrom:nil,},EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/bin/,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:6443,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:os-release,ReadOnly:false,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:system-cni-dir,ReadOnly:false,MountPath:/host/etc/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:multus-cni-dir,ReadOnly:false,MountPath:/host/var/run/multus/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:kube-api-access-zf2qr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,} start failed in pod multus-4xhw7_openshift-multus(84f9cd8a-5e58-4aff-aedd-906ed37dc97d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:12.309184 2588 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-4xhw7" podUID=84f9cd8a-5e58-4aff-aedd-906ed37dc97d Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.312642 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-nxzr8" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.312652 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.312953319Z" level=info msg="Got pod network &{Name:prometheus-adapter-5894cc86c7-qhmbh Namespace:openshift-monitoring ID:a68ba2c2702e0f26953b0481e42659c0abc11c4f57afaf50c0d6c186047f92d4 UID:a40bf20d-67a4-45e7-b701-ca5a75376c1f NetNS:/var/run/netns/0ef21f34-ce89-4cc1-b1bf-60bb857132a2 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.313035459Z" level=info msg="Adding pod openshift-monitoring_prometheus-adapter-5894cc86c7-qhmbh to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.313364 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-8xrbh" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.313686568Z" level=info msg="Running pod sandbox: openshift-ingress-canary/ingress-canary-8xrbh/POD" id=5c911e85-e00a-4fdc-a522-11a401d26bd1 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.313795847Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.313853272Z" level=info msg="Running pod sandbox: openshift-cluster-node-tuning-operator/tuned-smwj7/POD" id=6f835aa3-fbbe-4c5e-beba-473942bf0424 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.314009005Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.313686576Z" level=info msg="Running pod sandbox: openshift-dns/node-resolver-nxzr8/POD" id=fc534bdd-8af0-490d-a0cf-4458782d0181 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.314173388Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.323257 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-g48vj\" (UniqueName: \"kubernetes.io/projected/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-kube-api-access-g48vj\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.345639 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.346022163Z" level=info msg="Running pod sandbox: openshift-monitoring/prometheus-k8s-0/POD" id=71e3a947-76ff-4d18-a8dc-205bbf21124d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.346147544Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.347976 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.348338139Z" level=info msg="Running pod sandbox: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/POD" id=616dc127-1067-46be-98b2-28c19895d97c name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.348379414Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.348619033Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=6f835aa3-fbbe-4c5e-beba-473942bf0424 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.359084 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-xw6d2" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.362451527Z" level=info msg="Ran pod sandbox e0df29cbf3197efe5e4c221145d0f72e155c977f3d50c18c2f38862f24a5522e with infra container: openshift-cluster-node-tuning-operator/tuned-smwj7/POD" id=6f835aa3-fbbe-4c5e-beba-473942bf0424 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.363073002Z" level=info msg="Running pod sandbox: openshift-image-registry/node-ca-xw6d2/POD" id=9d068ee5-27fc-40f7-9268-5e0f9c36735a name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.363190159Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.364067639Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ca6ba83315a4882f5c6c4e01c79238e1b794b0e8bfdc2f004990f516e8e0b056" id=fa25dbe8-4bca-4347-90cf-a7a16a90cce7 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.371452177Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b2f8382287975a85853941eec2c606723c2ed1c24f59fa46ebdb470b91a481f5,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ca6ba83315a4882f5c6c4e01c79238e1b794b0e8bfdc2f004990f516e8e0b056],Size_:505526574,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=fa25dbe8-4bca-4347-90cf-a7a16a90cce7 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:12.374823 2588 kuberuntime_manager.go:862] container &Container{Name:tuned,Image:registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ca6ba83315a4882f5c6c4e01c79238e1b794b0e8bfdc2f004990f516e8e0b056,Command:[/var/lib/tuned/bin/run start],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:WATCH_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:OCP_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:RESYNC_PERIOD,Value:600,ValueFrom:nil,},EnvVar{Name:RELEASE_VERSION,Value:4.12.0-0.ci-2023-01-05-021459,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},},VolumeMounts:[]VolumeMount{VolumeMount{Name:var-lib-tuned-profiles-data,ReadOnly:false,MountPath:/var/lib/tuned/profiles-data,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:etc,ReadOnly:false,MountPath:/etc/modprobe.d,SubPath:modprobe.d,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:etc,ReadOnly:false,MountPath:/etc/sysconfig,SubPath:sysconfig,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:etc,ReadOnly:true,MountPath:/etc/sysctl.d,SubPath:sysctl.d,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:etc,ReadOnly:true,MountPath:/etc/sysctl.conf,SubPath:sysctl.conf,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:etc,ReadOnly:false,MountPath:/etc/systemd,SubPath:systemd,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:sys,ReadOnly:false,MountPath:/sys,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:var-run-dbus,ReadOnly:true,MountPath:/var/run/dbus,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:run-systemd-system,ReadOnly:true,MountPath:/run/systemd/system,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:lib-modules,ReadOnly:true,MountPath:/lib/modules,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:host,ReadOnly:false,MountPath:/host,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:kube-api-access-b7xz4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/var/lib/tuned/bin/run stop],},HTTPGet:nil,TCPSocket:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,} start failed in pod tuned-smwj7_openshift-cluster-node-tuning-operator(91265f01-9dcb-4046-961e-ad4c544224d7): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:12.375031 2588 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tuned\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-cluster-node-tuning-operator/tuned-smwj7" podUID=91265f01-9dcb-4046-961e-ad4c544224d7 Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.374842 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.375648407Z" level=info msg="Running pod sandbox: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/POD" id=c89c2627-bb06-4db8-b345-71d0fc5cb42c name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.375688902Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.381879858Z" level=info msg="Got pod network &{Name:ingress-canary-8xrbh Namespace:openshift-ingress-canary ID:bc4aa1a421b0c7450185bcf26f1acfb0471356c226b5d1eb5fea1eb8c214d725 UID:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748 NetNS:/var/run/netns/48587174-1261-4ae7-908c-76565765a732 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.381902978Z" level=info msg="Adding pod openshift-ingress-canary_ingress-canary-8xrbh to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.388448 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-2pfzf" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.391850327Z" level=info msg="Running pod sandbox: openshift-dns/dns-default-2pfzf/POD" id=e708422a-b746-4e79-8e2e-c00d1dc48840 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.392146102Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.393120774Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=fc534bdd-8af0-490d-a0cf-4458782d0181 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.397290 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.397632725Z" level=info msg="Running pod sandbox: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/POD" id=c92ba848-1321-41e5-ab61-e6a211aaa3ab name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.397753006Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:41:12.399495 2588 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f88cbe5_1221_4e6f_a6c9_f22da152b43f.slice/crio-91eb7fe6d0c596f220833b072fadd2715083a9931fdb3ed2bbf7986380dce490.scope WatchSource:0}: Error finding container 91eb7fe6d0c596f220833b072fadd2715083a9931fdb3ed2bbf7986380dce490: Status 404 returned error can't find the container with id 91eb7fe6d0c596f220833b072fadd2715083a9931fdb3ed2bbf7986380dce490 Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.409022270Z" level=info msg="Ran pod sandbox 91eb7fe6d0c596f220833b072fadd2715083a9931fdb3ed2bbf7986380dce490 with infra container: openshift-dns/node-resolver-nxzr8/POD" id=fc534bdd-8af0-490d-a0cf-4458782d0181 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.409772518Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f93a29f8d7455a5067ecde85394835bda877174193ab693b2b010d96c8cd9183" id=b9396caa-4f1a-4f01-96ec-1b06bae49fdc name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.410145154Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:9b1e063a0faa2683ff00ee1f8b324af985c1c142530a63c7f8729ec8dab01b38,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f93a29f8d7455a5067ecde85394835bda877174193ab693b2b010d96c8cd9183],Size_:480731528,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b9396caa-4f1a-4f01-96ec-1b06bae49fdc name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:12.410623 2588 kuberuntime_manager.go:862] container &Container{Name:dns-node-resolver,Image:registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f93a29f8d7455a5067ecde85394835bda877174193ab693b2b010d96c8cd9183,Command:[/bin/bash -c #!/bin/bash Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: set -uo pipefail Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: trap 'jobs -p | xargs kill || true; wait; exit 0' TERM Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: OPENSHIFT_MARKER="openshift-generated-node-resolver" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: HOSTS_FILE="/etc/hosts" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: TEMP_FILE="/etc/hosts.tmp" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: IFS=', ' read -r -a services <<< "${SERVICES}" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # Make a temporary file with the old hosts file's attributes. Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: cp -f --attributes-only "${HOSTS_FILE}" "${TEMP_FILE}" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: while true; do Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: declare -A svc_ips Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: for svc in "${services[@]}"; do Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # Fetch service IP from cluster dns if present. We make several tries Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # to do it: IPv4, IPv6, IPv4 over TCP and IPv6 over TCP. The two last ones Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # are for deployments with Kuryr on older OpenStack (OSP13) - those do not Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # support UDP loadbalancers and require reaching DNS through TCP. Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: cmds=('dig -t A @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: 'dig -t AAAA @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: 'dig -t A +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: 'dig -t AAAA +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"') Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: for i in ${!cmds[*]} Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: do Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: ips=($(eval "${cmds[i]}")) Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: if [[ "$?" -eq 0 && "${#ips[@]}" -ne 0 ]]; then Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: svc_ips["${svc}"]="${ips[@]}" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: break Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: fi Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: done Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: done Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # Update /etc/hosts only if we get valid service IPs Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # We will not update /etc/hosts when there is coredns service outage or api unavailability Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # Stale entries could exist in /etc/hosts if the service is deleted Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: if [[ -n "${svc_ips[*]-}" ]]; then Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # Build a new hosts file from /etc/hosts with our custom entries filtered out Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: grep -v "# ${OPENSHIFT_MARKER}" "${HOSTS_FILE}" > "${TEMP_FILE}" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # Append resolver entries for services Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: for svc in "${!svc_ips[@]}"; do Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: for ip in ${svc_ips[${svc}]}; do Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: echo "${ip} ${svc} ${svc}.${CLUSTER_DOMAIN} # ${OPENSHIFT_MARKER}" >> "${TEMP_FILE}" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: done Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: done Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # TODO: Update /etc/hosts atomically to avoid any inconsistent behavior Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # Replace /etc/hosts with our modified version if needed Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: cmp "${TEMP_FILE}" "${HOSTS_FILE}" || cp -f "${TEMP_FILE}" "${HOSTS_FILE}" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # TEMP_FILE is not removed to avoid file create/delete and attributes copy churn Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: fi Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: sleep 60 & wait Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: unset svc_ips Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: done Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:SERVICES,Value:image-registry.openshift-image-registry.svc,ValueFrom:nil,},EnvVar{Name:NAMESERVER,Value:172.30.0.10,ValueFrom:nil,},EnvVar{Name:CLUSTER_DOMAIN,Value:cluster.local,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{22020096 0} {} 21Mi BinarySI},},},VolumeMounts:[]VolumeMount{VolumeMount{Name:hosts-file,ReadOnly:false,MountPath:/etc/hosts,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:kube-api-access-g7www,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,} start failed in pod node-resolver-nxzr8_openshift-dns(8f88cbe5-1221-4e6f-a6c9-f22da152b43f): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:12.410683 2588 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dns-node-resolver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-dns/node-resolver-nxzr8" podUID=8f88cbe5-1221-4e6f-a6c9-f22da152b43f Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.417711477Z" level=info msg="Got pod network &{Name:thanos-querier-6f4f5fb7cf-rdm9l Namespace:openshift-monitoring ID:b9d09b59ecb8e2b530349c78a99778d6d8fc71efffc651d62a1e3fbb2005564d UID:7c63c9be-d1ad-44b6-8b33-09af3ee314af NetNS:/var/run/netns/7677b9cc-a641-408b-91a9-8419c8619347 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.417826073Z" level=info msg="Adding pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.427233867Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=616dc127-1067-46be-98b2-28c19895d97c name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.431380 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.434162252Z" level=info msg="Running pod sandbox: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/POD" id=fe43d99d-75da-4c33-84c4-0f91cfd3cfd4 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.434327941Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.442857 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.445446916Z" level=info msg="Running pod sandbox: openshift-ingress/router-default-5cf6bbdff9-tnffs/POD" id=d4ed3a90-2212-4ef9-8521-f3b9472cee6b name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.445499375Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:41:12.453379 2588 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1bb84a70_232c_4467_ac0b_647030155908.slice/crio-16cbcad68c73abf56080eb14225620b7f239901aad0b8916effb3d6a9dc1a771.scope WatchSource:0}: Error finding container 16cbcad68c73abf56080eb14225620b7f239901aad0b8916effb3d6a9dc1a771: Status 404 returned error can't find the container with id 16cbcad68c73abf56080eb14225620b7f239901aad0b8916effb3d6a9dc1a771 Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.458225828Z" level=info msg="Ran pod sandbox 16cbcad68c73abf56080eb14225620b7f239901aad0b8916effb3d6a9dc1a771 with infra container: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/POD" id=616dc127-1067-46be-98b2-28c19895d97c name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.459293002Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=9d068ee5-27fc-40f7-9268-5e0f9c36735a name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.460365159Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7" id=3d45ab14-07aa-485e-b35c-5f0464750fc6 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.460570706Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:6e78a44a973f2298da6f51c1418094403ee8c6c2b57bde9d3285b9e2ac2ba129,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7],Size_:397889589,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=3d45ab14-07aa-485e-b35c-5f0464750fc6 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:12.461052 2588 kuberuntime_manager.go:862] container &Container{Name:csi-driver,Image:registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7,Command:[],Args:[--driver=diskplugin.csi.alibabacloud.com --endpoint=$(CSI_ENDPOINT) --logtostderr --v=2],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:healthz,HostPort:10300,ContainerPort:10300,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:CSI_ENDPOINT,Value:unix://csi/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:MAX_VOLUMES_PERNODE,Value:15,ValueFrom:nil,},EnvVar{Name:SERVICE_TYPE,Value:plugin,ValueFrom:nil,},EnvVar{Name:DISK_AD_CONTROLLER,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{134217728 0} {} BinarySI},},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,},VolumeMount{Name:plugin-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:device-dir,ReadOnly:false,MountPath:/dev,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:non-standard-root-system-trust-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:kube-api-access-vzzxr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{1 0 healthz},Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:3,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:5,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,} start failed in pod alibaba-disk-csi-driver-node-5sqb4_openshift-cluster-csi-drivers(1bb84a70-232c-4467-ac0b-647030155908): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.461285994Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:72c764b55efc147864960930ae05b7928c5f976eb76820b38997f0df7afab820" id=b21bd82f-4bc3-40e5-9663-33beeea42d6e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.461982831Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:0bb939dc4e915f972f52c355cea411db098a9c605b2ca4bc1048447ae297391e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:72c764b55efc147864960930ae05b7928c5f976eb76820b38997f0df7afab820],Size_:293414029,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b21bd82f-4bc3-40e5-9663-33beeea42d6e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:12.462316 2588 kuberuntime_manager.go:862] container &Container{Name:csi-node-driver-registrar,Image:registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:72c764b55efc147864960930ae05b7928c5f976eb76820b38997f0df7afab820,Command:[],Args:[--csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH) --v=2],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/diskplugin.csi.alibabacloud.com/csi.sock,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugin-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:kube-api-access-vzzxr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/bin/sh -c rm -rf /registration/diskplugin.csi.alibabacloud.com.sock /csi/csi.sock],},HTTPGet:nil,TCPSocket:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,} start failed in pod alibaba-disk-csi-driver-node-5sqb4_openshift-cluster-csi-drivers(1bb84a70-232c-4467-ac0b-647030155908): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.462490308Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ea25c2f3749dee6981359b4655a9acfcd4d95547066ccdf328354b548da74932" id=1e2c6289-87b1-4d04-a448-eefff9815399 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.462637125Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:58a5591f9334b70902e27f9a172f3e141b8f07a879507b850a76c2499014eec9,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ea25c2f3749dee6981359b4655a9acfcd4d95547066ccdf328354b548da74932],Size_:291396124,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=1e2c6289-87b1-4d04-a448-eefff9815399 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:12.462864 2588 kuberuntime_manager.go:862] container &Container{Name:csi-liveness-probe,Image:registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ea25c2f3749dee6981359b4655a9acfcd4d95547066ccdf328354b548da74932,Command:[],Args:[--csi-address=/csi/csi.sock --probe-timeout=3s --health-port=10300 --v=2],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugin-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:kube-api-access-vzzxr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,} start failed in pod alibaba-disk-csi-driver-node-5sqb4_openshift-cluster-csi-drivers(1bb84a70-232c-4467-ac0b-647030155908): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.464446780Z" level=info msg="Got pod network &{Name:kube-state-metrics-6798b548b9-xl88q Namespace:openshift-monitoring ID:56657e8828648ef03ee57a604c40642f3fe11d7eaf9907e27b69fd6d2ac4609e UID:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e NetNS:/var/run/netns/13a07410-2a86-4abc-b7cb-dddc3b44117e Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:12.464478 2588 pod_workers.go:965] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"csi-driver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"csi-liveness-probe\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" podUID=1bb84a70-232c-4467-ac0b-647030155908 Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.464594809Z" level=info msg="Adding pod openshift-monitoring_kube-state-metrics-6798b548b9-xl88q to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.466621145Z" level=info msg="Got pod network &{Name:dns-default-2pfzf Namespace:openshift-dns ID:df8a985d595378c8d023c40bbf337dd18cbe474fb195d37d4cab28fb6e4e8e97 UID:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 NetNS:/var/run/netns/3bd81ca3-6cd1-42e1-b5af-cc2f9135e45d Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:41:12.469522 2588 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod659b8bb4_81b5_49bf_8127_1d326f30ba2f.slice/crio-93662ac8682d21720eb888cc865f999977edd4dcd1e585e0bae6c3aeeb2d9bba.scope WatchSource:0}: Error finding container 93662ac8682d21720eb888cc865f999977edd4dcd1e585e0bae6c3aeeb2d9bba: Status 404 returned error can't find the container with id 93662ac8682d21720eb888cc865f999977edd4dcd1e585e0bae6c3aeeb2d9bba Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.469566040Z" level=info msg="Adding pod openshift-dns_dns-default-2pfzf to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.473567600Z" level=info msg="Ran pod sandbox 93662ac8682d21720eb888cc865f999977edd4dcd1e585e0bae6c3aeeb2d9bba with infra container: openshift-image-registry/node-ca-xw6d2/POD" id=9d068ee5-27fc-40f7-9268-5e0f9c36735a name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.474468406Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8" id=1e7b5bc1-44ff-4901-975f-32b6f2fae28e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.478027094Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:c9e2954d8b9d9e1594d0a16b90086a9297ffd65773a6aa4a64f325e871ce83d1,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8],Size_:363009057,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=1e7b5bc1-44ff-4901-975f-32b6f2fae28e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:12.478534 2588 kuberuntime_manager.go:862] container &Container{Name:node-ca,Image:registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8,Command:[/bin/sh -c trap 'jobs -p | xargs -r kill; echo shutting down node-ca; exit 0' TERM Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: while [ true ]; Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: do Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: for f in $(ls /tmp/serviceca); do Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: echo $f Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: ca_file_path="/tmp/serviceca/${f}" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: f=$(echo $f | sed -r 's/(.*)\.\./\1:/') Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: reg_dir_path="/etc/docker/certs.d/${f}" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: if [ -e "${reg_dir_path}" ]; then Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: cp -u $ca_file_path $reg_dir_path/ca.crt Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: else Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: mkdir $reg_dir_path Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: cp $ca_file_path $reg_dir_path/ca.crt Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: fi Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: done Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: for d in $(ls /etc/docker/certs.d); do Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: echo $d Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: dp=$(echo $d | sed -r 's/(.*):/\1\.\./') Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: reg_conf_path="/tmp/serviceca/${dp}" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: if [ ! -e "${reg_conf_path}" ]; then Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: rm -rf /etc/docker/certs.d/$d Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: fi Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: done Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: sleep 60 & wait ${!} Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: done Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{10485760 0} {} 10Mi BinarySI},},},VolumeMounts:[]VolumeMount{VolumeMount{Name:serviceca,ReadOnly:false,MountPath:/tmp/serviceca,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:host,ReadOnly:false,MountPath:/etc/docker/certs.d,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:kube-api-access-h4sjn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*1001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,} start failed in pod node-ca-xw6d2_openshift-image-registry(659b8bb4-81b5-49bf-8127-1d326f30ba2f): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:12.478591 2588 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"node-ca\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-image-registry/node-ca-xw6d2" podUID=659b8bb4-81b5-49bf-8127-1d326f30ba2f Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.504153790Z" level=info msg="Got pod network &{Name:openshift-state-metrics-6f5dcdf79b-wq256 Namespace:openshift-monitoring ID:f8841eb8a77afd8023660e94ffd508c50ea7025758849f56e786914d6fd4c8d1 UID:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 NetNS:/var/run/netns/82c20a6e-750c-4278-84a9-395a5a2ea52b Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.504423659Z" level=info msg="Adding pod openshift-monitoring_openshift-state-metrics-6f5dcdf79b-wq256 to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.506111471Z" level=info msg="Got pod network &{Name:router-default-5cf6bbdff9-tnffs Namespace:openshift-ingress ID:3307dc0da12fd52b662f10d53f8d87884254715041942b38a02e5c3023a51b72 UID:c3b10ff4-a243-4a4e-b27d-f613e258666d NetNS:/var/run/netns/f01414fc-9125-434c-b723-d4b97fb30ecb Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.506135024Z" level=info msg="Adding pod openshift-ingress_router-default-5cf6bbdff9-tnffs to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:12.584967 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.585434655Z" level=info msg="Running pod sandbox: openshift-network-diagnostics/network-check-target-xmq2g/POD" id=66686923-ed64-4f02-af98-ba9095064cea name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.585491584Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.598309841Z" level=info msg="Got pod network &{Name:network-check-target-xmq2g Namespace:openshift-network-diagnostics ID:116a8242496b3ef676a5a564cd538412df8f6e3800ed923ea68d2687007a6d72 UID:f364a949-09d7-4cfb-83ff-e532b822a557 NetNS:/var/run/netns/88a729ff-9dec-4eae-9c39-4d6236f2267e Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:41:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:12.598336985Z" level=info msg="Adding pod openshift-network-diagnostics_network-check-target-xmq2g to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:13.041650 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-sdn/sdn-66nsp" event=&{ID:d5b0f5e6-3d70-4315-b9c5-1403036f517b Type:ContainerStarted Data:06a6d838dd853ce187ccf364676fd4e36976cae02f20a98aeda4b59df6df2d93} Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:13.045809 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-2r6nf" event=&{ID:aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f Type:ContainerStarted Data:acd08c77f0076092e020bedcc20e2c94faab21df9b7d5516eba6f0dbe370935b} Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:13.046964313Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:30387b6f2e605bb8d018e0c09f9d0275048fc5e7966137fbe6bd4307c88d3fdd" id=b53c4687-e5a3-4f26-9c24-77a9dcbc186a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:13.047253079Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:cff39d57d2c205a584f8bf4299b1e125fa1a96ac92f680117ba82c20334754d1,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:30387b6f2e605bb8d018e0c09f9d0275048fc5e7966137fbe6bd4307c88d3fdd],Size_:701241664,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b53c4687-e5a3-4f26-9c24-77a9dcbc186a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:13.050102 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-xw6d2" event=&{ID:659b8bb4-81b5-49bf-8127-1d326f30ba2f Type:ContainerStarted Data:93662ac8682d21720eb888cc865f999977edd4dcd1e585e0bae6c3aeeb2d9bba} Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:13.048798 2588 kuberuntime_manager.go:862] container &Container{Name:sdn,Image:registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:30387b6f2e605bb8d018e0c09f9d0275048fc5e7966137fbe6bd4307c88d3fdd,Command:[/bin/bash -c #!/bin/bash Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: set -euo pipefail Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # if another process is listening on the cni-server socket, wait until it exits Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: trap 'kill $(jobs -p); rm -f /etc/cni/net.d/80-openshift-network.conf ; exit 0' TERM Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: retries=0 Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: while true; do Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: if echo 'test' | socat - UNIX-CONNECT:/var/run/openshift-sdn/cniserver/socket &>/dev/null; then Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: echo "warning: Another process is currently listening on the CNI socket, waiting 15s ..." 2>&1 Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: sleep 15 & wait Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: (( retries += 1 )) Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: else Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: break Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: fi Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: if [[ "${retries}" -gt 40 ]]; then Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: echo "error: Another process is currently listening on the CNI socket, exiting" 2>&1 Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: exit 1 Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: fi Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: done Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # local environment overrides Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: if [[ -f /etc/sysconfig/openshift-sdn ]]; then Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: set -o allexport Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: source /etc/sysconfig/openshift-sdn Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: set +o allexport Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: fi Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: #BUG: cdc accidentally mounted /etc/sysconfig/openshift-sdn as DirectoryOrCreate; clean it up so we can ultimately mount /etc/sysconfig/openshift-sdn as FileOrCreate Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # Once this is released, then we can mount it properly Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: if [[ -d /etc/sysconfig/openshift-sdn ]]; then Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: rmdir /etc/sysconfig/openshift-sdn || true Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: fi Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # configmap-based overrides Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: if [[ -f /env/${K8S_NODE_NAME} ]]; then Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: set -o allexport Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: source /env/${K8S_NODE_NAME} Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: set +o allexport Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: fi Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # Take over network functions on the node Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: rm -f /etc/cni/net.d/80-openshift-network.conf Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: cp -f /opt/cni/bin/openshift-sdn /host-cni-bin/ Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: mtu_override_flag= Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: if [[ -f /config-mtu-migration/mtu.yaml ]]; then Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: mtu_override_flag="--mtu-override /config-mtu-migration/mtu.yaml" Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: fi Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # Launch the network process Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: exec /usr/bin/openshift-sdn-node \ Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: --node-name ${K8S_NODE_NAME} --node-ip ${K8S_NODE_IP} \ Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: --platform-type AlibabaCloud \ Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: --proxy-config /config/kube-proxy-config.yaml \ Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: ${mtu_override_flag} \ Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: --v ${OPENSHIFT_SDN_LOG_LEVEL:-2} Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:healthz,HostPort:10256,ContainerPort:10256,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:6443,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com,ValueFrom:nil,},EnvVar{Name:OPENSHIFT_DNS_DOMAIN,Value:cluster.local,ValueFrom:nil,},EnvVar{Name:K8S_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:K8S_NODE_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.hostIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{209715200 0} {} BinarySI},},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/config,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:config-mtu-migration,ReadOnly:true,MountPath:/config-mtu-migration,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:host-var-run-crio,ReadOnly:false,MountPath:/var/run/crio,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:host-var-run-dbus,ReadOnly:true,MountPath:/var/run/dbus/,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:host-var-run-ovs,ReadOnly:true,MountPath:/var/run/openvswitch/,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:host-var-run-kubernetes,ReadOnly:true,MountPath:/var/run/kubernetes/,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:host-run-netns,ReadOnly:true,MountPath:/run/netns,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,},VolumeMount{Name:host-var-run-openshift-sdn,ReadOnly:false,MountPath:/var/run/openshift-sdn,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,},VolumeMount{Name:host-cni-bin,ReadOnly:false,MountPath:/host-cni-bin,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:host-cni-conf,ReadOnly:false,MountPath:/etc/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:host-var-lib-cni-networks-openshift-sdn,ReadOnly:false,MountPath:/var/lib/cni/networks/openshift-sdn,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:host-modules,ReadOnly:true,MountPath:/lib/modules,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:etc-sysconfig,ReadOnly:true,MountPath:/etc/sysconfig,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:kube-api-access-8nd4t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[test -f /etc/cni/net.d/80-openshift-network.conf],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[rm -f /etc/cni/net.d/80-openshift-network.conf /host-cni-bin/openshift-sdn],},HTTPGet:nil,TCPSocket:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,} start failed in pod sdn-66nsp_openshift-sdn(d5b0f5e6-3d70-4315-b9c5-1403036f517b): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:13.052789926Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=b5b756c8-6fee-43f7-990a-e7c22c611dbf name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:13.053024332Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b5b756c8-6fee-43f7-990a-e7c22c611dbf name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:13.054296709Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180" id=4549c175-190f-453b-bea8-ed93f4eaa7af name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:13.054557956Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b4cde205857c1e0a984b38c61275a8370136432c7c84dbfbd7f4835f26bbb935,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180],Size_:297268865,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=4549c175-190f-453b-bea8-ed93f4eaa7af name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:13.055470624Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8" id=049c93dd-3c54-494f-92b3-3ea312fceb0e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:13.055726759Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:c9e2954d8b9d9e1594d0a16b90086a9297ffd65773a6aa4a64f325e871ce83d1,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8],Size_:363009057,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=049c93dd-3c54-494f-92b3-3ea312fceb0e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:13.056524 2588 kuberuntime_manager.go:862] init container &Container{Name:init-textfile,Image:registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180,Command:[/bin/sh -c [[ ! -d /node_exporter/collectors/init ]] || find /node_exporter/collectors/init -perm /111 -type f -exec {} \;],Args:[],WorkingDir:/var/node_exporter/textfile,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:TMPDIR,Value:/tmp,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{1 -3} {} 1m DecimalSI},memory: {{1048576 0} {} 1Mi BinarySI},},},VolumeMounts:[]VolumeMount{VolumeMount{Name:node-exporter-textfile,ReadOnly:false,MountPath:/var/node_exporter/textfile,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:node-exporter-wtmp,ReadOnly:true,MountPath:/var/log/wtmp,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:kube-api-access-dfs2p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,} start failed in pod node-exporter-2r6nf_openshift-monitoring(aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:13.056687 2588 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init-textfile\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-monitoring/node-exporter-2r6nf" podUID=aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:13.057084 2588 kuberuntime_manager.go:862] container &Container{Name:kube-rbac-proxy,Image:registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9,Command:[/bin/bash -c #!/bin/bash Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: set -euo pipefail Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: TLS_PK=/etc/pki/tls/metrics-certs/tls.key Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: TLS_CERT=/etc/pki/tls/metrics-certs/tls.crt Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # As the secret mount is optional we must wait for the files to be present. Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # The service is created in monitor.yaml and this is created in sdn.yaml. Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # If it isn't created there is probably an issue so we want to crashloop. Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: TS=$(date +%s) Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: WARN_TS=$(( ${TS} + $(( 20 * 60)) )) Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: HAS_LOGGED_INFO=0 Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: log_missing_certs(){ Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: CUR_TS=$(date +%s) Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: if [[ "${CUR_TS}" -gt "WARN_TS" ]]; then Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: echo $(date -Iseconds) WARN: sdn-metrics-certs not mounted after 20 minutes. Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: elif [[ "${HAS_LOGGED_INFO}" -eq 0 ]] ; then Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: echo $(date -Iseconds) INFO: sdn-metrics-certs not mounted. Waiting 20 minutes. Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: HAS_LOGGED_INFO=1 Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: fi Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: } Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: while [[ ! -f "${TLS_PK}" || ! -f "${TLS_CERT}" ]] ; do Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: log_missing_certs Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: sleep 5 Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: done Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: echo $(date -Iseconds) INFO: sdn-metrics-certs mounted, starting kube-rbac-proxy Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: exec /usr/bin/kube-rbac-proxy \ Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: --logtostderr \ Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: --secure-listen-address=:9101 \ Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 \ Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: --upstream=http://127.0.0.1:29101/ \ Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: --tls-private-key-file=${TLS_PK} \ Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: --tls-cert-file=${TLS_CERT} Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:9101,ContainerPort:9101,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{20971520 0} {} 20Mi BinarySI},},},VolumeMounts:[]VolumeMount{VolumeMount{Name:sdn-metrics-certs,ReadOnly:true,MountPath:/etc/pki/tls/metrics-certs,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:kube-api-access-8nd4t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,} start failed in pod sdn-66nsp_openshift-sdn(d5b0f5e6-3d70-4315-b9c5-1403036f517b): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:13.059608 2588 pod_workers.go:965] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"sdn\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-sdn/sdn-66nsp" podUID=d5b0f5e6-3d70-4315-b9c5-1403036f517b Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:13.057533 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerStarted Data:e9053289b0b05a28538d40135f5851bcbb4fef2e500409bd4a7c787c89dc9515} Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:13.059453 2588 kuberuntime_manager.go:862] container &Container{Name:node-ca,Image:registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8,Command:[/bin/sh -c trap 'jobs -p | xargs -r kill; echo shutting down node-ca; exit 0' TERM Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: while [ true ]; Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: do Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: for f in $(ls /tmp/serviceca); do Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: echo $f Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: ca_file_path="/tmp/serviceca/${f}" Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: f=$(echo $f | sed -r 's/(.*)\.\./\1:/') Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: reg_dir_path="/etc/docker/certs.d/${f}" Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: if [ -e "${reg_dir_path}" ]; then Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: cp -u $ca_file_path $reg_dir_path/ca.crt Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: else Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: mkdir $reg_dir_path Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: cp $ca_file_path $reg_dir_path/ca.crt Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: fi Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: done Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: for d in $(ls /etc/docker/certs.d); do Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: echo $d Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: dp=$(echo $d | sed -r 's/(.*):/\1\.\./') Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: reg_conf_path="/tmp/serviceca/${dp}" Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: if [ ! -e "${reg_conf_path}" ]; then Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: rm -rf /etc/docker/certs.d/$d Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: fi Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: done Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: sleep 60 & wait ${!} Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: done Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{10485760 0} {} 10Mi BinarySI},},},VolumeMounts:[]VolumeMount{VolumeMount{Name:serviceca,ReadOnly:false,MountPath:/tmp/serviceca,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:host,ReadOnly:false,MountPath:/etc/docker/certs.d,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:kube-api-access-h4sjn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*1001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,} start failed in pod node-ca-xw6d2_openshift-image-registry(659b8bb4-81b5-49bf-8127-1d326f30ba2f): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:13.061102 2588 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"node-ca\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-image-registry/node-ca-xw6d2" podUID=659b8bb4-81b5-49bf-8127-1d326f30ba2f Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:13.062024630Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:3fc008580e4e404960b2de4020a2939ab6421bdf8f51c0b60af4ad903272c741" id=d05a296d-886b-41b6-bbc2-e2e1858e8dc8 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:13.062325183Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:6bb2a44a1f0506c18ef471cac7f7e8991d1e6b9d649ccbce50f751dd2863d06d,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:3fc008580e4e404960b2de4020a2939ab6421bdf8f51c0b60af4ad903272c741],Size_:382411122,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=d05a296d-886b-41b6-bbc2-e2e1858e8dc8 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:13.061321 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" event=&{ID:1bb84a70-232c-4467-ac0b-647030155908 Type:ContainerStarted Data:16cbcad68c73abf56080eb14225620b7f239901aad0b8916effb3d6a9dc1a771} Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:13.071538 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" event=&{ID:997b7ab4-939e-465c-9c7d-4a2ebe3a797a Type:ContainerStarted Data:ae18be6d794af60a14ce35ff3392698b0b7e426a5128047f31455ade1f2b79dd} Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:13.072820719Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7" id=b54cee35-d19e-4f99-b8e4-3f700dfa45ff name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:13.073104987Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:6e78a44a973f2298da6f51c1418094403ee8c6c2b57bde9d3285b9e2ac2ba129,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7],Size_:397889589,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b54cee35-d19e-4f99-b8e4-3f700dfa45ff name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:13.064413 2588 kuberuntime_manager.go:862] init container &Container{Name:egress-router-binary-copy,Image:registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:3fc008580e4e404960b2de4020a2939ab6421bdf8f51c0b60af4ad903272c741,Command:[/entrypoint/cnibincopy.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL7_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel7/bin/,ValueFrom:nil,},EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/bin/,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:os-release,ReadOnly:true,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:kube-api-access-wq4k7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,} start failed in pod multus-additional-cni-plugins-5fgxc_openshift-multus(c0e47249-a5ff-460e-958d-53577387c154): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:13.077376 2588 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"egress-router-binary-copy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" podUID=c0e47249-a5ff-460e-958d-53577387c154 Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:13.077587 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-nxzr8" event=&{ID:8f88cbe5-1221-4e6f-a6c9-f22da152b43f Type:ContainerStarted Data:91eb7fe6d0c596f220833b072fadd2715083a9931fdb3ed2bbf7986380dce490} Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:13.078159 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-smwj7" event=&{ID:91265f01-9dcb-4046-961e-ad4c544224d7 Type:ContainerStarted Data:e0df29cbf3197efe5e4c221145d0f72e155c977f3d50c18c2f38862f24a5522e} Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:13.078322 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4xhw7" event=&{ID:84f9cd8a-5e58-4aff-aedd-906ed37dc97d Type:ContainerStarted Data:bb5b6562ae64906e3450c53a90bf4df6ea55af8ce39868944ba2cdcf7eab1188} Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:13.077903 2588 kuberuntime_manager.go:862] container &Container{Name:csi-driver,Image:registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7,Command:[],Args:[--driver=diskplugin.csi.alibabacloud.com --endpoint=$(CSI_ENDPOINT) --logtostderr --v=2],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:healthz,HostPort:10300,ContainerPort:10300,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:CSI_ENDPOINT,Value:unix://csi/csi.sock,ValueFrom:nil,},EnvVar{Name:KUBE_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:MAX_VOLUMES_PERNODE,Value:15,ValueFrom:nil,},EnvVar{Name:SERVICE_TYPE,Value:plugin,ValueFrom:nil,},EnvVar{Name:DISK_AD_CONTROLLER,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{134217728 0} {} BinarySI},},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kubelet-dir,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*Bidirectional,SubPathExpr:,},VolumeMount{Name:plugin-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:device-dir,ReadOnly:false,MountPath:/dev,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:non-standard-root-system-trust-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:kube-api-access-vzzxr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{1 0 healthz},Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:3,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:5,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,} start failed in pod alibaba-disk-csi-driver-node-5sqb4_openshift-cluster-csi-drivers(1bb84a70-232c-4467-ac0b-647030155908): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:13.079838117Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cb9edf2251410c340a6125ff331924ab57a72ad3166bd9e1305b6d3030284990" id=57d915c1-c7f4-4c4a-b153-0e89b3912c08 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:13.080051617Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2b208acf1f002c143fdb248dbee2517f7b6d168def4cfef1f3188939123d0dc3,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cb9edf2251410c340a6125ff331924ab57a72ad3166bd9e1305b6d3030284990],Size_:536350945,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=57d915c1-c7f4-4c4a-b153-0e89b3912c08 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:13.080130970Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:72c764b55efc147864960930ae05b7928c5f976eb76820b38997f0df7afab820" id=dde268de-9b9d-4436-ad75-5cbc7f4df183 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:13.080545590Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:0bb939dc4e915f972f52c355cea411db098a9c605b2ca4bc1048447ae297391e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:72c764b55efc147864960930ae05b7928c5f976eb76820b38997f0df7afab820],Size_:293414029,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=dde268de-9b9d-4436-ad75-5cbc7f4df183 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:13.081028 2588 kuberuntime_manager.go:862] container &Container{Name:csi-node-driver-registrar,Image:registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:72c764b55efc147864960930ae05b7928c5f976eb76820b38997f0df7afab820,Command:[],Args:[--csi-address=$(ADDRESS) --kubelet-registration-path=$(DRIVER_REG_SOCK_PATH) --v=2],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADDRESS,Value:/csi/csi.sock,ValueFrom:nil,},EnvVar{Name:DRIVER_REG_SOCK_PATH,Value:/var/lib/kubelet/plugins/diskplugin.csi.alibabacloud.com/csi.sock,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugin-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:registration-dir,ReadOnly:false,MountPath:/registration,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:kube-api-access-vzzxr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/bin/sh -c rm -rf /registration/diskplugin.csi.alibabacloud.com.sock /csi/csi.sock],},HTTPGet:nil,TCPSocket:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,} start failed in pod alibaba-disk-csi-driver-node-5sqb4_openshift-cluster-csi-drivers(1bb84a70-232c-4467-ac0b-647030155908): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:13.081920072Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ea25c2f3749dee6981359b4655a9acfcd4d95547066ccdf328354b548da74932" id=4b9070e3-122c-43fb-8ecb-17d40435d676 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:13.082126052Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:58a5591f9334b70902e27f9a172f3e141b8f07a879507b850a76c2499014eec9,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ea25c2f3749dee6981359b4655a9acfcd4d95547066ccdf328354b548da74932],Size_:291396124,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=4b9070e3-122c-43fb-8ecb-17d40435d676 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:13.083170 2588 kuberuntime_manager.go:862] container &Container{Name:machine-config-daemon,Image:registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cb9edf2251410c340a6125ff331924ab57a72ad3166bd9e1305b6d3030284990,Command:[/usr/bin/machine-config-daemon],Args:[start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},},VolumeMounts:[]VolumeMount{VolumeMount{Name:rootfs,ReadOnly:false,MountPath:/rootfs,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:kube-api-access-5xrks,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,} start failed in pod machine-config-daemon-hf8f5_openshift-machine-config-operator(997b7ab4-939e-465c-9c7d-4a2ebe3a797a): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:13.085343024Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f93a29f8d7455a5067ecde85394835bda877174193ab693b2b010d96c8cd9183" id=d6e68c90-08e4-4c4f-9144-8e5c89069247 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:13.085550161Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:9b1e063a0faa2683ff00ee1f8b324af985c1c142530a63c7f8729ec8dab01b38,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f93a29f8d7455a5067ecde85394835bda877174193ab693b2b010d96c8cd9183],Size_:480731528,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=d6e68c90-08e4-4c4f-9144-8e5c89069247 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:13.085634488Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=4c462dea-5960-41d3-843c-81482725deee name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:13.085798978Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=4c462dea-5960-41d3-843c-81482725deee name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:13.087360651Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ca6ba83315a4882f5c6c4e01c79238e1b794b0e8bfdc2f004990f516e8e0b056" id=fdd5b4be-112f-4c27-adb9-a9e706911f6e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:13.087556189Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b2f8382287975a85853941eec2c606723c2ed1c24f59fa46ebdb470b91a481f5,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ca6ba83315a4882f5c6c4e01c79238e1b794b0e8bfdc2f004990f516e8e0b056],Size_:505526574,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=fdd5b4be-112f-4c27-adb9-a9e706911f6e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:13.089307554Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=98558171-ae20-47d3-a7d8-3db1aad1e644 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:13.089556762Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=98558171-ae20-47d3-a7d8-3db1aad1e644 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:13.090348 2588 kuberuntime_manager.go:862] container &Container{Name:csi-liveness-probe,Image:registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ea25c2f3749dee6981359b4655a9acfcd4d95547066ccdf328354b548da74932,Command:[],Args:[--csi-address=/csi/csi.sock --probe-timeout=3s --health-port=10300 --v=2],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugin-dir,ReadOnly:false,MountPath:/csi,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:kube-api-access-vzzxr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,} start failed in pod alibaba-disk-csi-driver-node-5sqb4_openshift-cluster-csi-drivers(1bb84a70-232c-4467-ac0b-647030155908): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:13.092769 2588 pod_workers.go:965] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"csi-driver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"csi-node-driver-registrar\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"csi-liveness-probe\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" podUID=1bb84a70-232c-4467-ac0b-647030155908 Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:13.092242 2588 kuberuntime_manager.go:862] container &Container{Name:tuned,Image:registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ca6ba83315a4882f5c6c4e01c79238e1b794b0e8bfdc2f004990f516e8e0b056,Command:[/var/lib/tuned/bin/run start],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:WATCH_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:OCP_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:RESYNC_PERIOD,Value:600,ValueFrom:nil,},EnvVar{Name:RELEASE_VERSION,Value:4.12.0-0.ci-2023-01-05-021459,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},},VolumeMounts:[]VolumeMount{VolumeMount{Name:var-lib-tuned-profiles-data,ReadOnly:false,MountPath:/var/lib/tuned/profiles-data,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:etc,ReadOnly:false,MountPath:/etc/modprobe.d,SubPath:modprobe.d,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:etc,ReadOnly:false,MountPath:/etc/sysconfig,SubPath:sysconfig,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:etc,ReadOnly:true,MountPath:/etc/sysctl.d,SubPath:sysctl.d,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:etc,ReadOnly:true,MountPath:/etc/sysctl.conf,SubPath:sysctl.conf,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:etc,ReadOnly:false,MountPath:/etc/systemd,SubPath:systemd,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:sys,ReadOnly:false,MountPath:/sys,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:var-run-dbus,ReadOnly:true,MountPath:/var/run/dbus,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:run-systemd-system,ReadOnly:true,MountPath:/run/systemd/system,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:lib-modules,ReadOnly:true,MountPath:/lib/modules,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:host,ReadOnly:false,MountPath:/host,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:kube-api-access-b7xz4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:&Lifecycle{PostStart:nil,PreStop:&LifecycleHandler{Exec:&ExecAction{Command:[/var/lib/tuned/bin/run stop],},HTTPGet:nil,TCPSocket:nil,},},TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,} start failed in pod tuned-smwj7_openshift-cluster-node-tuning-operator(91265f01-9dcb-4046-961e-ad4c544224d7): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:13.092021 2588 kuberuntime_manager.go:862] container &Container{Name:dns-node-resolver,Image:registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f93a29f8d7455a5067ecde85394835bda877174193ab693b2b010d96c8cd9183,Command:[/bin/bash -c #!/bin/bash Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: set -uo pipefail Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: trap 'jobs -p | xargs kill || true; wait; exit 0' TERM Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: OPENSHIFT_MARKER="openshift-generated-node-resolver" Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: HOSTS_FILE="/etc/hosts" Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: TEMP_FILE="/etc/hosts.tmp" Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: IFS=', ' read -r -a services <<< "${SERVICES}" Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # Make a temporary file with the old hosts file's attributes. Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: cp -f --attributes-only "${HOSTS_FILE}" "${TEMP_FILE}" Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: while true; do Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: declare -A svc_ips Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: for svc in "${services[@]}"; do Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # Fetch service IP from cluster dns if present. We make several tries Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # to do it: IPv4, IPv6, IPv4 over TCP and IPv6 over TCP. The two last ones Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # are for deployments with Kuryr on older OpenStack (OSP13) - those do not Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # support UDP loadbalancers and require reaching DNS through TCP. Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: cmds=('dig -t A @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: 'dig -t AAAA @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: 'dig -t A +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: 'dig -t AAAA +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"') Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: for i in ${!cmds[*]} Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: do Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: ips=($(eval "${cmds[i]}")) Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: if [[ "$?" -eq 0 && "${#ips[@]}" -ne 0 ]]; then Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: svc_ips["${svc}"]="${ips[@]}" Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: break Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: fi Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: done Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: done Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # Update /etc/hosts only if we get valid service IPs Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # We will not update /etc/hosts when there is coredns service outage or api unavailability Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # Stale entries could exist in /etc/hosts if the service is deleted Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: if [[ -n "${svc_ips[*]-}" ]]; then Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # Build a new hosts file from /etc/hosts with our custom entries filtered out Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: grep -v "# ${OPENSHIFT_MARKER}" "${HOSTS_FILE}" > "${TEMP_FILE}" Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # Append resolver entries for services Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: for svc in "${!svc_ips[@]}"; do Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: for ip in ${svc_ips[${svc}]}; do Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: echo "${ip} ${svc} ${svc}.${CLUSTER_DOMAIN} # ${OPENSHIFT_MARKER}" >> "${TEMP_FILE}" Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: done Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: done Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # TODO: Update /etc/hosts atomically to avoid any inconsistent behavior Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # Replace /etc/hosts with our modified version if needed Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: cmp "${TEMP_FILE}" "${HOSTS_FILE}" || cp -f "${TEMP_FILE}" "${HOSTS_FILE}" Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: # TEMP_FILE is not removed to avoid file create/delete and attributes copy churn Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: fi Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: sleep 60 & wait Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: unset svc_ips Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: done Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:SERVICES,Value:image-registry.openshift-image-registry.svc,ValueFrom:nil,},EnvVar{Name:NAMESERVER,Value:172.30.0.10,ValueFrom:nil,},EnvVar{Name:CLUSTER_DOMAIN,Value:cluster.local,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{22020096 0} {} 21Mi BinarySI},},},VolumeMounts:[]VolumeMount{VolumeMount{Name:hosts-file,ReadOnly:false,MountPath:/etc/hosts,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:kube-api-access-g7www,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,} start failed in pod node-resolver-nxzr8_openshift-dns(8f88cbe5-1221-4e6f-a6c9-f22da152b43f): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:13.094457 2588 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tuned\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-cluster-node-tuning-operator/tuned-smwj7" podUID=91265f01-9dcb-4046-961e-ad4c544224d7 Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:13.097048 2588 kuberuntime_manager.go:862] container &Container{Name:kube-multus,Image:registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f,Command:[/bin/bash -ec --],Args:[/entrypoint/cnibincopy.sh; exec /entrypoint.sh --multus-conf-file=auto --multus-autoconfig-dir=/host/var/run/multus/cni/net.d --multus-kubeconfig-file-host=/etc/kubernetes/cni/net.d/multus.d/multus.kubeconfig --readiness-indicator-file=/var/run/multus/cni/net.d/80-openshift-network.conf --cleanup-config-on-exit=true --namespace-isolation=true --multus-log-level=verbose --cni-version=0.3.1 --additional-bin-dir=/opt/multus/bin --skip-multus-binary-copy=true - "--global-namespaces=default,openshift-multus,openshift-sriov-network-operator" Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: ],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL7_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel7/bin/,ValueFrom:nil,},EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/bin/,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:6443,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:api-int.pliu-dev-alicloud.alicloud-qe.devcluster.openshift.com,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:os-release,ReadOnly:false,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:system-cni-dir,ReadOnly:false,MountPath:/host/etc/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:multus-cni-dir,ReadOnly:false,MountPath:/host/var/run/multus/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:kube-api-access-zf2qr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,} start failed in pod multus-4xhw7_openshift-multus(84f9cd8a-5e58-4aff-aedd-906ed37dc97d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:13.101094 2588 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-4xhw7" podUID=84f9cd8a-5e58-4aff-aedd-906ed37dc97d Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:13.097752 2588 kuberuntime_manager.go:862] container &Container{Name:oauth-proxy,Image:registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b,Command:[],Args:[--https-address=:9001 --provider=openshift --openshift-service-account=machine-config-daemon --upstream=http://127.0.0.1:8797 --tls-cert=/etc/tls/private/tls.crt --tls-key=/etc/tls/private/tls.key --cookie-secret-file=/etc/tls/cookie-secret/cookie-secret --openshift-sar={"resource": "namespaces", "verb": "get"} --openshift-delegate-urls={"/": {"resource": "namespaces", "verb": "get"}}],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:9001,ContainerPort:9001,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},},VolumeMounts:[]VolumeMount{VolumeMount{Name:proxy-tls,ReadOnly:false,MountPath:/etc/tls/private,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:cookie-secret,ReadOnly:false,MountPath:/etc/tls/cookie-secret,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:kube-api-access-5xrks,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,} start failed in pod machine-config-daemon-hf8f5_openshift-machine-config-operator(997b7ab4-939e-465c-9c7d-4a2ebe3a797a): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:13.098023 2588 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dns-node-resolver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-dns/node-resolver-nxzr8" podUID=8f88cbe5-1221-4e6f-a6c9-f22da152b43f Jan 05 09:41:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:13.102935 2588 pod_workers.go:965] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"machine-config-daemon\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"oauth-proxy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" podUID=997b7ab4-939e-465c-9c7d-4a2ebe3a797a Jan 05 09:41:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:23.793501075Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=03607f67-7dcf-445b-aa9c-0b7ba011536d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:23.793781638Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=03607f67-7dcf-445b-aa9c-0b7ba011536d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:23.793829717Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f93a29f8d7455a5067ecde85394835bda877174193ab693b2b010d96c8cd9183" id=18b306a1-8204-447f-ad49-4affd7d3ef8e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:23.793914673Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cb9edf2251410c340a6125ff331924ab57a72ad3166bd9e1305b6d3030284990" id=8a61e046-33dc-4da1-9ce2-53fd0b0cd555 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:23.794011720Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:9b1e063a0faa2683ff00ee1f8b324af985c1c142530a63c7f8729ec8dab01b38,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f93a29f8d7455a5067ecde85394835bda877174193ab693b2b010d96c8cd9183],Size_:480731528,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=18b306a1-8204-447f-ad49-4affd7d3ef8e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:23.794098721Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2b208acf1f002c143fdb248dbee2517f7b6d168def4cfef1f3188939123d0dc3,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cb9edf2251410c340a6125ff331924ab57a72ad3166bd9e1305b6d3030284990],Size_:536350945,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=8a61e046-33dc-4da1-9ce2-53fd0b0cd555 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:23.795060415Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cb9edf2251410c340a6125ff331924ab57a72ad3166bd9e1305b6d3030284990" id=bde21e14-093e-4dc9-a618-468cf0faef33 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:23.795404561Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2b208acf1f002c143fdb248dbee2517f7b6d168def4cfef1f3188939123d0dc3,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cb9edf2251410c340a6125ff331924ab57a72ad3166bd9e1305b6d3030284990],Size_:536350945,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=bde21e14-093e-4dc9-a618-468cf0faef33 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:23.795729380Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f93a29f8d7455a5067ecde85394835bda877174193ab693b2b010d96c8cd9183" id=9b1cfd0c-188b-40ec-b483-6a540b651b3c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:23.795962855Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:9b1e063a0faa2683ff00ee1f8b324af985c1c142530a63c7f8729ec8dab01b38,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f93a29f8d7455a5067ecde85394835bda877174193ab693b2b010d96c8cd9183],Size_:480731528,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=9b1cfd0c-188b-40ec-b483-6a540b651b3c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:23.796072028Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=a0463cd4-7a7a-4a2e-bacf-847eaf825ff7 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:23.796311643Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=a0463cd4-7a7a-4a2e-bacf-847eaf825ff7 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:23.796883464Z" level=info msg="Creating container: openshift-machine-config-operator/machine-config-daemon-hf8f5/machine-config-daemon" id=cdbef542-25f3-4ec4-9e91-836c7a854d09 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:23.797286349Z" level=info msg="Creating container: openshift-multus/multus-4xhw7/kube-multus" id=94edb259-e45f-4fe8-b49f-17e7aa6d692b name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:23.797533078Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:23.797595912Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:23.798370191Z" level=info msg="Creating container: openshift-dns/node-resolver-nxzr8/dns-node-resolver" id=762b4680-943c-4c7f-b062-7b76db0a8d1c name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:23.798476435Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:23.995651450Z" level=info msg="Created container a101bd9b2fa9f40fbb8f57ada55ba85a2b32e0a1b4b3095834cce4f256a57a6d: openshift-machine-config-operator/machine-config-daemon-hf8f5/machine-config-daemon" id=cdbef542-25f3-4ec4-9e91-836c7a854d09 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:23.996965669Z" level=info msg="Starting container: a101bd9b2fa9f40fbb8f57ada55ba85a2b32e0a1b4b3095834cce4f256a57a6d" id=df9d624f-53d7-4643-ba18-0644e32084e7 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:41:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:24.012689236Z" level=info msg="Started container" PID=2883 containerID=a101bd9b2fa9f40fbb8f57ada55ba85a2b32e0a1b4b3095834cce4f256a57a6d description=openshift-machine-config-operator/machine-config-daemon-hf8f5/machine-config-daemon id=df9d624f-53d7-4643-ba18-0644e32084e7 name=/runtime.v1.RuntimeService/StartContainer sandboxID=ae18be6d794af60a14ce35ff3392698b0b7e426a5128047f31455ade1f2b79dd Jan 05 09:41:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:24.045522057Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=bd25e9fa-b6f2-4861-a610-b9e83c8d973c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:24.045747404Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=bd25e9fa-b6f2-4861-a610-b9e83c8d973c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:24.047115622Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=434f583a-c30c-430b-8c82-78f98404398c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:24.047432696Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=434f583a-c30c-430b-8c82-78f98404398c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:24.049017606Z" level=info msg="Creating container: openshift-machine-config-operator/machine-config-daemon-hf8f5/oauth-proxy" id=51c23b1c-5a26-4840-a65c-bc29d795aaed name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:24.049149400Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:24.101143 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" event=&{ID:997b7ab4-939e-465c-9c7d-4a2ebe3a797a Type:ContainerStarted Data:a101bd9b2fa9f40fbb8f57ada55ba85a2b32e0a1b4b3095834cce4f256a57a6d} Jan 05 09:41:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:24.203885829Z" level=info msg="Created container b3efd856773bdf7155e86d05756f3b9cedcd2cf5f1f70db09bd557c4076feebb: openshift-machine-config-operator/machine-config-daemon-hf8f5/oauth-proxy" id=51c23b1c-5a26-4840-a65c-bc29d795aaed name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:24.205838078Z" level=info msg="Starting container: b3efd856773bdf7155e86d05756f3b9cedcd2cf5f1f70db09bd557c4076feebb" id=f231a070-d336-47c2-8b6b-e5c765c5f90e name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:41:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:24.236256715Z" level=info msg="Started container" PID=2922 containerID=b3efd856773bdf7155e86d05756f3b9cedcd2cf5f1f70db09bd557c4076feebb description=openshift-machine-config-operator/machine-config-daemon-hf8f5/oauth-proxy id=f231a070-d336-47c2-8b6b-e5c765c5f90e name=/runtime.v1.RuntimeService/StartContainer sandboxID=ae18be6d794af60a14ce35ff3392698b0b7e426a5128047f31455ade1f2b79dd Jan 05 09:41:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:24.629723250Z" level=info msg="Created container dacbcea32fcb08b92bce49a0d7ac9968dc159622dde6d950a43fc195447b8b66: openshift-multus/multus-4xhw7/kube-multus" id=94edb259-e45f-4fe8-b49f-17e7aa6d692b name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:24.631772416Z" level=info msg="Starting container: dacbcea32fcb08b92bce49a0d7ac9968dc159622dde6d950a43fc195447b8b66" id=f19108b7-d1b2-40a9-ab17-3996a1528da5 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:41:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:24.653337181Z" level=info msg="Started container" PID=2975 containerID=dacbcea32fcb08b92bce49a0d7ac9968dc159622dde6d950a43fc195447b8b66 description=openshift-multus/multus-4xhw7/kube-multus id=f19108b7-d1b2-40a9-ab17-3996a1528da5 name=/runtime.v1.RuntimeService/StartContainer sandboxID=bb5b6562ae64906e3450c53a90bf4df6ea55af8ce39868944ba2cdcf7eab1188 Jan 05 09:41:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:24.675603657Z" level=info msg="Created container 077ba6e68a89d52a0235be6b670ac7de4696698f28533f3907f3fb962f717756: openshift-dns/node-resolver-nxzr8/dns-node-resolver" id=762b4680-943c-4c7f-b062-7b76db0a8d1c name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:24.675981429Z" level=info msg="Starting container: 077ba6e68a89d52a0235be6b670ac7de4696698f28533f3907f3fb962f717756" id=6a4cc11a-4e92-4268-8c94-22b31f3f255c name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:41:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:24.684476388Z" level=info msg="Started container" PID=3033 containerID=077ba6e68a89d52a0235be6b670ac7de4696698f28533f3907f3fb962f717756 description=openshift-dns/node-resolver-nxzr8/dns-node-resolver id=6a4cc11a-4e92-4268-8c94-22b31f3f255c name=/runtime.v1.RuntimeService/StartContainer sandboxID=91eb7fe6d0c596f220833b072fadd2715083a9931fdb3ed2bbf7986380dce490 Jan 05 09:41:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:24.686888662Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_1fd7b117-1b82-4957-9f7d-ce20e4e5f61b\"" Jan 05 09:41:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:24.734416178Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:24.734454449Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:24.796429680Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8" id=afeef50b-8ff4-49fc-ad40-4b21ab8c4ebb name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:24.796675857Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:c9e2954d8b9d9e1594d0a16b90086a9297ffd65773a6aa4a64f325e871ce83d1,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8],Size_:363009057,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=afeef50b-8ff4-49fc-ad40-4b21ab8c4ebb name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:24.801814845Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8" id=53c007f3-28f5-4aa8-b5a7-71feb50069c7 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:24.802045619Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:c9e2954d8b9d9e1594d0a16b90086a9297ffd65773a6aa4a64f325e871ce83d1,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8],Size_:363009057,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=53c007f3-28f5-4aa8-b5a7-71feb50069c7 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:24.807722145Z" level=info msg="Creating container: openshift-image-registry/node-ca-xw6d2/node-ca" id=6351fb99-0ee7-4970-b191-9b3f442c7022 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:24.807867604Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:24.826762712Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/multus\"" Jan 05 09:41:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:24.928283463Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:24.928312918Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:24.928347649Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_1fd7b117-1b82-4957-9f7d-ce20e4e5f61b\"" Jan 05 09:41:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:25.111279 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-nxzr8" event=&{ID:8f88cbe5-1221-4e6f-a6c9-f22da152b43f Type:ContainerStarted Data:077ba6e68a89d52a0235be6b670ac7de4696698f28533f3907f3fb962f717756} Jan 05 09:41:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:25.117521 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" event=&{ID:997b7ab4-939e-465c-9c7d-4a2ebe3a797a Type:ContainerStarted Data:b3efd856773bdf7155e86d05756f3b9cedcd2cf5f1f70db09bd557c4076feebb} Jan 05 09:41:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:25.120133 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4xhw7" event=&{ID:84f9cd8a-5e58-4aff-aedd-906ed37dc97d Type:ContainerStarted Data:dacbcea32fcb08b92bce49a0d7ac9968dc159622dde6d950a43fc195447b8b66} Jan 05 09:41:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:25.202246215Z" level=info msg="Created container 519bf3596b61bb638a20f8b10fb133f44e6c68d1e0dbabd3ae375fd1b2a55104: openshift-image-registry/node-ca-xw6d2/node-ca" id=6351fb99-0ee7-4970-b191-9b3f442c7022 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:25.203684796Z" level=info msg="Starting container: 519bf3596b61bb638a20f8b10fb133f44e6c68d1e0dbabd3ae375fd1b2a55104" id=e9a0b601-6269-46d0-8cba-3801d4f29a6d name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:41:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:25.215025795Z" level=info msg="Started container" PID=3227 containerID=519bf3596b61bb638a20f8b10fb133f44e6c68d1e0dbabd3ae375fd1b2a55104 description=openshift-image-registry/node-ca-xw6d2/node-ca id=e9a0b601-6269-46d0-8cba-3801d4f29a6d name=/runtime.v1.RuntimeService/StartContainer sandboxID=93662ac8682d21720eb888cc865f999977edd4dcd1e585e0bae6c3aeeb2d9bba Jan 05 09:41:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:25.793898599Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7" id=7e7a5e57-7dee-4c26-b7f6-0b10136d432c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:25.794364492Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:6e78a44a973f2298da6f51c1418094403ee8c6c2b57bde9d3285b9e2ac2ba129,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7],Size_:397889589,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=7e7a5e57-7dee-4c26-b7f6-0b10136d432c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:25.795589887Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7" id=bc5c4602-be7b-4153-9fd6-e6fdea0380de name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:25.795756716Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:6e78a44a973f2298da6f51c1418094403ee8c6c2b57bde9d3285b9e2ac2ba129,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7],Size_:397889589,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=bc5c4602-be7b-4153-9fd6-e6fdea0380de name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:25.796429987Z" level=info msg="Creating container: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-driver" id=de13548d-68b4-4cc0-881e-c6b3c33392bc name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:25.796551073Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:25.888173184Z" level=info msg="Created container 67cdebe351c893eaa507d5db4eacca0b7a2988004a12dc2f5ac95212a0847d23: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-driver" id=de13548d-68b4-4cc0-881e-c6b3c33392bc name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:25.889018672Z" level=info msg="Starting container: 67cdebe351c893eaa507d5db4eacca0b7a2988004a12dc2f5ac95212a0847d23" id=282e6dcb-ff97-4093-bca9-059c1c72089a name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:41:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:25.897010388Z" level=info msg="Started container" PID=3290 containerID=67cdebe351c893eaa507d5db4eacca0b7a2988004a12dc2f5ac95212a0847d23 description=openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-driver id=282e6dcb-ff97-4093-bca9-059c1c72089a name=/runtime.v1.RuntimeService/StartContainer sandboxID=16cbcad68c73abf56080eb14225620b7f239901aad0b8916effb3d6a9dc1a771 Jan 05 09:41:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:25.905059463Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:72c764b55efc147864960930ae05b7928c5f976eb76820b38997f0df7afab820" id=2e172446-ece9-4fd1-952d-fb6c531f6a52 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:25.905455013Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:0bb939dc4e915f972f52c355cea411db098a9c605b2ca4bc1048447ae297391e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:72c764b55efc147864960930ae05b7928c5f976eb76820b38997f0df7afab820],Size_:293414029,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=2e172446-ece9-4fd1-952d-fb6c531f6a52 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:25.906223818Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:72c764b55efc147864960930ae05b7928c5f976eb76820b38997f0df7afab820" id=1a00e177-195f-49c0-b20c-047be6f4e15e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:25.906428811Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:0bb939dc4e915f972f52c355cea411db098a9c605b2ca4bc1048447ae297391e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:72c764b55efc147864960930ae05b7928c5f976eb76820b38997f0df7afab820],Size_:293414029,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=1a00e177-195f-49c0-b20c-047be6f4e15e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:25.907112088Z" level=info msg="Creating container: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-node-driver-registrar" id=32dd051b-2e0b-4aac-a0cb-3a5570bf08d2 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:25 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:25.907243911Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.000853504Z" level=info msg="Created container a5d628e10517a9c71ea3c3fcecbba33834d81b193fb3d83cf5f536aa4c7af861: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-node-driver-registrar" id=32dd051b-2e0b-4aac-a0cb-3a5570bf08d2 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.001421421Z" level=info msg="Starting container: a5d628e10517a9c71ea3c3fcecbba33834d81b193fb3d83cf5f536aa4c7af861" id=9efc2301-5ad9-48b7-b6c6-8586438953bd name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.010708138Z" level=info msg="Started container" PID=3326 containerID=a5d628e10517a9c71ea3c3fcecbba33834d81b193fb3d83cf5f536aa4c7af861 description=openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-node-driver-registrar id=9efc2301-5ad9-48b7-b6c6-8586438953bd name=/runtime.v1.RuntimeService/StartContainer sandboxID=16cbcad68c73abf56080eb14225620b7f239901aad0b8916effb3d6a9dc1a771 Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.024714987Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ea25c2f3749dee6981359b4655a9acfcd4d95547066ccdf328354b548da74932" id=264a93f5-bf9a-4bd8-959a-1df6a197d500 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.025186300Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:58a5591f9334b70902e27f9a172f3e141b8f07a879507b850a76c2499014eec9,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ea25c2f3749dee6981359b4655a9acfcd4d95547066ccdf328354b548da74932],Size_:291396124,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=264a93f5-bf9a-4bd8-959a-1df6a197d500 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.029452652Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ea25c2f3749dee6981359b4655a9acfcd4d95547066ccdf328354b548da74932" id=6db0b191-4b0a-4b3d-ba6f-5560468f266c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.029645500Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:58a5591f9334b70902e27f9a172f3e141b8f07a879507b850a76c2499014eec9,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ea25c2f3749dee6981359b4655a9acfcd4d95547066ccdf328354b548da74932],Size_:291396124,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=6db0b191-4b0a-4b3d-ba6f-5560468f266c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.036552436Z" level=info msg="Creating container: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-liveness-probe" id=d673f967-58b1-4262-9d41-058f18501b65 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.036760912Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:26.123667 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" event=&{ID:1bb84a70-232c-4467-ac0b-647030155908 Type:ContainerStarted Data:a5d628e10517a9c71ea3c3fcecbba33834d81b193fb3d83cf5f536aa4c7af861} Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:26.123846 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" event=&{ID:1bb84a70-232c-4467-ac0b-647030155908 Type:ContainerStarted Data:67cdebe351c893eaa507d5db4eacca0b7a2988004a12dc2f5ac95212a0847d23} Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:26.124936 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-xw6d2" event=&{ID:659b8bb4-81b5-49bf-8127-1d326f30ba2f Type:ContainerStarted Data:519bf3596b61bb638a20f8b10fb133f44e6c68d1e0dbabd3ae375fd1b2a55104} Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.151970199Z" level=info msg="Created container d5f8f3cbb8f4dc651cb6f673e2779934b25961eaad0ae6a146a877b33d7a04b2: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-liveness-probe" id=d673f967-58b1-4262-9d41-058f18501b65 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.152317297Z" level=info msg="Starting container: d5f8f3cbb8f4dc651cb6f673e2779934b25961eaad0ae6a146a877b33d7a04b2" id=2720dc36-11d0-4240-b517-4cd2f66c077f name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.162179615Z" level=info msg="Started container" PID=3372 containerID=d5f8f3cbb8f4dc651cb6f673e2779934b25961eaad0ae6a146a877b33d7a04b2 description=openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-liveness-probe id=2720dc36-11d0-4240-b517-4cd2f66c077f name=/runtime.v1.RuntimeService/StartContainer sandboxID=16cbcad68c73abf56080eb14225620b7f239901aad0b8916effb3d6a9dc1a771 Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.794126529Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:30387b6f2e605bb8d018e0c09f9d0275048fc5e7966137fbe6bd4307c88d3fdd" id=52fa899e-386b-47f4-acad-71b12263c445 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.794704271Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:cff39d57d2c205a584f8bf4299b1e125fa1a96ac92f680117ba82c20334754d1,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:30387b6f2e605bb8d018e0c09f9d0275048fc5e7966137fbe6bd4307c88d3fdd],Size_:701241664,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=52fa899e-386b-47f4-acad-71b12263c445 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.795622771Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ca6ba83315a4882f5c6c4e01c79238e1b794b0e8bfdc2f004990f516e8e0b056" id=9e2c9ef3-ceee-4af1-848a-19a48f4c9c7b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.795852695Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b2f8382287975a85853941eec2c606723c2ed1c24f59fa46ebdb470b91a481f5,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ca6ba83315a4882f5c6c4e01c79238e1b794b0e8bfdc2f004990f516e8e0b056],Size_:505526574,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=9e2c9ef3-ceee-4af1-848a-19a48f4c9c7b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.796657039Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:30387b6f2e605bb8d018e0c09f9d0275048fc5e7966137fbe6bd4307c88d3fdd" id=b8bf6be4-181b-4cd0-a1ce-db0c99464463 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.796923903Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:cff39d57d2c205a584f8bf4299b1e125fa1a96ac92f680117ba82c20334754d1,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:30387b6f2e605bb8d018e0c09f9d0275048fc5e7966137fbe6bd4307c88d3fdd],Size_:701241664,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b8bf6be4-181b-4cd0-a1ce-db0c99464463 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.798115778Z" level=info msg="Creating container: openshift-sdn/sdn-66nsp/sdn" id=a144f78b-46df-4962-baef-dfb4560008cd name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.798243097Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.829673690Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ca6ba83315a4882f5c6c4e01c79238e1b794b0e8bfdc2f004990f516e8e0b056" id=40033bf0-2552-4312-b985-f897f00c00f6 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.829860840Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b2f8382287975a85853941eec2c606723c2ed1c24f59fa46ebdb470b91a481f5,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ca6ba83315a4882f5c6c4e01c79238e1b794b0e8bfdc2f004990f516e8e0b056],Size_:505526574,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=40033bf0-2552-4312-b985-f897f00c00f6 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.830967957Z" level=info msg="Creating container: openshift-cluster-node-tuning-operator/tuned-smwj7/tuned" id=c1108845-695d-46a3-a936-e06707c7badd name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.831089222Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.947532487Z" level=info msg="Created container 19e17c3e6276c0974e5ea7cee47cf21ef28e4f76c9800d7f245956a5a30930ae: openshift-sdn/sdn-66nsp/sdn" id=a144f78b-46df-4962-baef-dfb4560008cd name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.948755227Z" level=info msg="Starting container: 19e17c3e6276c0974e5ea7cee47cf21ef28e4f76c9800d7f245956a5a30930ae" id=ab75dffb-0931-422a-8639-196a974962b5 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.959790839Z" level=info msg="Started container" PID=3436 containerID=19e17c3e6276c0974e5ea7cee47cf21ef28e4f76c9800d7f245956a5a30930ae description=openshift-sdn/sdn-66nsp/sdn id=ab75dffb-0931-422a-8639-196a974962b5 name=/runtime.v1.RuntimeService/StartContainer sandboxID=06a6d838dd853ce187ccf364676fd4e36976cae02f20a98aeda4b59df6df2d93 Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.972834805Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=95acebb7-5ab4-4f53-b7b6-34126b22cc74 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.973085577Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=95acebb7-5ab4-4f53-b7b6-34126b22cc74 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.973835859Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=f79c8bed-34a7-4b02-b9ca-f43632e7cc71 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.974032480Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=f79c8bed-34a7-4b02-b9ca-f43632e7cc71 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.974880157Z" level=info msg="Creating container: openshift-sdn/sdn-66nsp/kube-rbac-proxy" id=594d5d5f-9afe-469f-b657-e0d989179000 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.974993275Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:26.995366730Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 09:41:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:27.039765327Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:27.039966701Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:27.040075163Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 09:41:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:27.069453994Z" level=info msg="Created container 13d74ea448697cb8f3566002c0b4971ff98d65f23ae68e8d201d09d31eb3eb4b: openshift-cluster-node-tuning-operator/tuned-smwj7/tuned" id=c1108845-695d-46a3-a936-e06707c7badd name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:27.070668256Z" level=info msg="Starting container: 13d74ea448697cb8f3566002c0b4971ff98d65f23ae68e8d201d09d31eb3eb4b" id=62237592-4cba-4d4f-8852-398678ec7978 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:41:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:27.086127561Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:27.086160799Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:27.086180336Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 09:41:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:27.099022438Z" level=info msg="Started container" PID=3464 containerID=13d74ea448697cb8f3566002c0b4971ff98d65f23ae68e8d201d09d31eb3eb4b description=openshift-cluster-node-tuning-operator/tuned-smwj7/tuned id=62237592-4cba-4d4f-8852-398678ec7978 name=/runtime.v1.RuntimeService/StartContainer sandboxID=e0df29cbf3197efe5e4c221145d0f72e155c977f3d50c18c2f38862f24a5522e Jan 05 09:41:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:27.110272020Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:27.110320781Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:27.110340108Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 09:41:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:27.128314292Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:27.128451439Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:27.130676 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" event=&{ID:1bb84a70-232c-4467-ac0b-647030155908 Type:ContainerStarted Data:d5f8f3cbb8f4dc651cb6f673e2779934b25961eaad0ae6a146a877b33d7a04b2} Jan 05 09:41:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:27.134307 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-sdn/sdn-66nsp" event=&{ID:d5b0f5e6-3d70-4315-b9c5-1403036f517b Type:ContainerStarted Data:19e17c3e6276c0974e5ea7cee47cf21ef28e4f76c9800d7f245956a5a30930ae} Jan 05 09:41:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:27.137345 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-smwj7" event=&{ID:91265f01-9dcb-4046-961e-ad4c544224d7 Type:ContainerStarted Data:13d74ea448697cb8f3566002c0b4971ff98d65f23ae68e8d201d09d31eb3eb4b} Jan 05 09:41:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:27.149599132Z" level=info msg="Created container ba950aa1e5aebd2e717e5f9c201e1ccfd4d70438d1b6d0f48fc608e5b8353acb: openshift-sdn/sdn-66nsp/kube-rbac-proxy" id=594d5d5f-9afe-469f-b657-e0d989179000 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:27.150098783Z" level=info msg="Starting container: ba950aa1e5aebd2e717e5f9c201e1ccfd4d70438d1b6d0f48fc608e5b8353acb" id=5ac04ba5-f006-4ff3-9b1d-31c359394349 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:41:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:27.162891938Z" level=info msg="Started container" PID=3517 containerID=ba950aa1e5aebd2e717e5f9c201e1ccfd4d70438d1b6d0f48fc608e5b8353acb description=openshift-sdn/sdn-66nsp/kube-rbac-proxy id=5ac04ba5-f006-4ff3-9b1d-31c359394349 name=/runtime.v1.RuntimeService/StartContainer sandboxID=06a6d838dd853ce187ccf364676fd4e36976cae02f20a98aeda4b59df6df2d93 Jan 05 09:41:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:27.794418188Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:3fc008580e4e404960b2de4020a2939ab6421bdf8f51c0b60af4ad903272c741" id=753b66cd-dda9-4d20-a9c8-824353fc9fbf name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:27.794760233Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:6bb2a44a1f0506c18ef471cac7f7e8991d1e6b9d649ccbce50f751dd2863d06d,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:3fc008580e4e404960b2de4020a2939ab6421bdf8f51c0b60af4ad903272c741],Size_:382411122,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=753b66cd-dda9-4d20-a9c8-824353fc9fbf name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:27.802583026Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:3fc008580e4e404960b2de4020a2939ab6421bdf8f51c0b60af4ad903272c741" id=ba97e58b-21f7-440a-b73d-3fbefdab4b6f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:27.802868633Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:6bb2a44a1f0506c18ef471cac7f7e8991d1e6b9d649ccbce50f751dd2863d06d,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:3fc008580e4e404960b2de4020a2939ab6421bdf8f51c0b60af4ad903272c741],Size_:382411122,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=ba97e58b-21f7-440a-b73d-3fbefdab4b6f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:27.803888748Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/egress-router-binary-copy" id=4fd46c4b-54f3-4682-826b-df6bd16162ca name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:27.804014794Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:28.142954 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-sdn/sdn-66nsp" event=&{ID:d5b0f5e6-3d70-4315-b9c5-1403036f517b Type:ContainerStarted Data:ba950aa1e5aebd2e717e5f9c201e1ccfd4d70438d1b6d0f48fc608e5b8353acb} Jan 05 09:41:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:28.143245 2588 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:28.168603758Z" level=info msg="Created container d174a8931ce09972e4f41a2f5881eda4aad39b7f793bcf994a20c7acc3f2f748: openshift-multus/multus-additional-cni-plugins-5fgxc/egress-router-binary-copy" id=4fd46c4b-54f3-4682-826b-df6bd16162ca name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:28.169009689Z" level=info msg="Starting container: d174a8931ce09972e4f41a2f5881eda4aad39b7f793bcf994a20c7acc3f2f748" id=80e245d0-3550-43f7-a39f-6a1deba12c5e name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:41:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:28.191251016Z" level=info msg="Started container" PID=3720 containerID=d174a8931ce09972e4f41a2f5881eda4aad39b7f793bcf994a20c7acc3f2f748 description=openshift-multus/multus-additional-cni-plugins-5fgxc/egress-router-binary-copy id=80e245d0-3550-43f7-a39f-6a1deba12c5e name=/runtime.v1.RuntimeService/StartContainer sandboxID=e9053289b0b05a28538d40135f5851bcbb4fef2e500409bd4a7c787c89dc9515 Jan 05 09:41:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:28.209915363Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_e2f5eff5-69d8-4794-a338-ce3c0af933e1\"" Jan 05 09:41:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:28.229974372Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:28.230005612Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:28.409962782Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/egress-router\"" Jan 05 09:41:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:28.428786613Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:28.428825087Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:28.430659303Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_e2f5eff5-69d8-4794-a338-ce3c0af933e1\"" Jan 05 09:41:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:28.793519295Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180" id=e2022975-2649-4e3d-9e0c-261bd7314b6b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:28.793757995Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b4cde205857c1e0a984b38c61275a8370136432c7c84dbfbd7f4835f26bbb935,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180],Size_:297268865,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=e2022975-2649-4e3d-9e0c-261bd7314b6b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:28.794826589Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180" id=3d3992ea-27de-4bf6-9671-1c757e6fb393 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:28.795042284Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b4cde205857c1e0a984b38c61275a8370136432c7c84dbfbd7f4835f26bbb935,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180],Size_:297268865,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=3d3992ea-27de-4bf6-9671-1c757e6fb393 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:28.796131639Z" level=info msg="Creating container: openshift-monitoring/node-exporter-2r6nf/init-textfile" id=ac615002-9ebf-472a-8e38-d775f1151315 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:28.796260041Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:28.889384288Z" level=info msg="Created container 9f6cdbb728dedef8960bc5f34adf71312fcf29d1e7c534217a674bd2f40ac0a4: openshift-monitoring/node-exporter-2r6nf/init-textfile" id=ac615002-9ebf-472a-8e38-d775f1151315 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:28.889778489Z" level=info msg="Starting container: 9f6cdbb728dedef8960bc5f34adf71312fcf29d1e7c534217a674bd2f40ac0a4" id=117ff201-a78a-4bf5-8d0c-06437ffec0b9 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:41:28 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:28.897145768Z" level=info msg="Started container" PID=3887 containerID=9f6cdbb728dedef8960bc5f34adf71312fcf29d1e7c534217a674bd2f40ac0a4 description=openshift-monitoring/node-exporter-2r6nf/init-textfile id=117ff201-a78a-4bf5-8d0c-06437ffec0b9 name=/runtime.v1.RuntimeService/StartContainer sandboxID=acd08c77f0076092e020bedcc20e2c94faab21df9b7d5516eba6f0dbe370935b Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.108346520Z" level=error msg="Failed to cleanup (probably retrying): failed to destroy network for pod sandbox k8s_prometheus-k8s-0_openshift-monitoring_4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879_0(04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0): error removing pod openshift-monitoring_prometheus-k8s-0 from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-monitoring/prometheus-k8s-0/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.109163667Z" level=info msg="Got pod network &{Name:prometheus-k8s-0 Namespace:openshift-monitoring ID:a108e4a07ee46c33fed5d4508ec6bf5478eacc0dc17a5ac7b7e9bfbd9fa620d3 UID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 NetNS:/var/run/netns/7b579e74-f3d2-4d6d-990a-eb17dda53649 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.109221735Z" level=info msg="Adding pod openshift-monitoring_prometheus-k8s-0 to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:29.143834 2588 generic.go:296] "Generic (PLEG): container finished" podID=aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f containerID="9f6cdbb728dedef8960bc5f34adf71312fcf29d1e7c534217a674bd2f40ac0a4" exitCode=0 Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:29.144126 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-2r6nf" event=&{ID:aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f Type:ContainerDied Data:9f6cdbb728dedef8960bc5f34adf71312fcf29d1e7c534217a674bd2f40ac0a4} Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.146041079Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180" id=c1591ca5-171f-4c42-b5a5-8eec79cb3bed name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.146489400Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b4cde205857c1e0a984b38c61275a8370136432c7c84dbfbd7f4835f26bbb935,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180],Size_:297268865,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=c1591ca5-171f-4c42-b5a5-8eec79cb3bed name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:29.147049 2588 generic.go:296] "Generic (PLEG): container finished" podID=c0e47249-a5ff-460e-958d-53577387c154 containerID="d174a8931ce09972e4f41a2f5881eda4aad39b7f793bcf994a20c7acc3f2f748" exitCode=0 Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:29.149223 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerDied Data:d174a8931ce09972e4f41a2f5881eda4aad39b7f793bcf994a20c7acc3f2f748} Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.150224182Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d5fc161170fcf68900b2694039803a2e7f1b31eb913566139a41f91851f166d7" id=4070a825-fde1-46dc-9643-374e3ad7b789 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.150535879Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:852cc803756e640141d6c3eb3e6607f2b0f7076ea77f9190b1f490288f40b3d8,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d5fc161170fcf68900b2694039803a2e7f1b31eb913566139a41f91851f166d7],Size_:574516705,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=4070a825-fde1-46dc-9643-374e3ad7b789 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.151517338Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180" id=b7d6deac-0495-4fbc-9137-345eaad83dd7 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.151709485Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b4cde205857c1e0a984b38c61275a8370136432c7c84dbfbd7f4835f26bbb935,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180],Size_:297268865,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b7d6deac-0495-4fbc-9137-345eaad83dd7 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.151898341Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d5fc161170fcf68900b2694039803a2e7f1b31eb913566139a41f91851f166d7" id=93425330-d3d1-4eab-8874-0ffabb335cd8 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.152077141Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:852cc803756e640141d6c3eb3e6607f2b0f7076ea77f9190b1f490288f40b3d8,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d5fc161170fcf68900b2694039803a2e7f1b31eb913566139a41f91851f166d7],Size_:574516705,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=93425330-d3d1-4eab-8874-0ffabb335cd8 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.153715292Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/cni-plugins" id=058c07a7-e880-4ba6-85db-ef89d697ad68 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.153836565Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.156381861Z" level=info msg="Creating container: openshift-monitoring/node-exporter-2r6nf/node-exporter" id=3928db2e-c3ff-4b94-8cb7-676b54cac1c2 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.156482941Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.360171078Z" level=info msg="CNI monitoring event REMOVE \"/etc/kubernetes/cni/net.d/00-multus.conf\"" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.415392105Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.415418058Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.415436602Z" level=info msg="CNI monitoring event CREATE \"/etc/kubernetes/cni/net.d/00-multus.conf\"" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.463124214Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.463353549Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.463459502Z" level=info msg="CNI monitoring event WRITE \"/etc/kubernetes/cni/net.d/00-multus.conf\"" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.492544323Z" level=info msg="Created container 53913239c5995dae15c3ae1ad17601117935102c6cb10665278cf79610336127: openshift-monitoring/node-exporter-2r6nf/node-exporter" id=3928db2e-c3ff-4b94-8cb7-676b54cac1c2 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.493047769Z" level=info msg="Starting container: 53913239c5995dae15c3ae1ad17601117935102c6cb10665278cf79610336127" id=c69c04fe-c5d8-4e4a-a295-d50220e7a437 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.512364270Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.512399816Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.512418235Z" level=info msg="CNI monitoring event CHMOD \"/etc/kubernetes/cni/net.d/00-multus.conf\"" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.517983839Z" level=info msg="Started container" PID=4044 containerID=53913239c5995dae15c3ae1ad17601117935102c6cb10665278cf79610336127 description=openshift-monitoring/node-exporter-2r6nf/node-exporter id=c69c04fe-c5d8-4e4a-a295-d50220e7a437 name=/runtime.v1.RuntimeService/StartContainer sandboxID=acd08c77f0076092e020bedcc20e2c94faab21df9b7d5516eba6f0dbe370935b Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.539948388Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=00558854-913c-4e63-b17a-82936cc7adeb name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.540173728Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=00558854-913c-4e63-b17a-82936cc7adeb name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.540992688Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=bae1dbae-cd7f-4bf3-8dda-9575e83a73c5 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.541191992Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=bae1dbae-cd7f-4bf3-8dda-9575e83a73c5 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.542328696Z" level=info msg="Creating container: openshift-monitoring/node-exporter-2r6nf/kube-rbac-proxy" id=d8c267f5-c70b-4c1a-b7a2-f717cc13dee8 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.542456263Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.669050199Z" level=info msg="Created container db32d5f1c7edaafe4b12f6a178251fd11f7647122a044136c8d596c85418df5c: openshift-multus/multus-additional-cni-plugins-5fgxc/cni-plugins" id=058c07a7-e880-4ba6-85db-ef89d697ad68 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.669582064Z" level=info msg="Starting container: db32d5f1c7edaafe4b12f6a178251fd11f7647122a044136c8d596c85418df5c" id=db70dfd2-25f0-4d24-8eeb-3daab55d5fd6 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.688780874Z" level=info msg="Started container" PID=4132 containerID=db32d5f1c7edaafe4b12f6a178251fd11f7647122a044136c8d596c85418df5c description=openshift-multus/multus-additional-cni-plugins-5fgxc/cni-plugins id=db70dfd2-25f0-4d24-8eeb-3daab55d5fd6 name=/runtime.v1.RuntimeService/StartContainer sandboxID=e9053289b0b05a28538d40135f5851bcbb4fef2e500409bd4a7c787c89dc9515 Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.711744438Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_54e566b9-4267-42da-923a-4d7bcd088cef\"" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.725488627Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.725520927Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.742643925Z" level=info msg="Created container be49f87e3a886d52fabc779de277793ff0cb222138bdd9809b5d437a60835b88: openshift-monitoring/node-exporter-2r6nf/kube-rbac-proxy" id=d8c267f5-c70b-4c1a-b7a2-f717cc13dee8 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.743113960Z" level=info msg="Starting container: be49f87e3a886d52fabc779de277793ff0cb222138bdd9809b5d437a60835b88" id=5d29131b-e23b-45a6-8704-0e5367c44567 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.756076268Z" level=info msg="Started container" PID=4165 containerID=be49f87e3a886d52fabc779de277793ff0cb222138bdd9809b5d437a60835b88 description=openshift-monitoring/node-exporter-2r6nf/kube-rbac-proxy id=5d29131b-e23b-45a6-8704-0e5367c44567 name=/runtime.v1.RuntimeService/StartContainer sandboxID=acd08c77f0076092e020bedcc20e2c94faab21df9b7d5516eba6f0dbe370935b Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.895071603Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/bandwidth\"" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.908637692Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.908669976Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.908688892Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/bridge\"" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.920048576Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.920082015Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.920099897Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/dhcp\"" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.935479440Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.935508638Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.935524604Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/firewall\"" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.948346231Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.948375736Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.948391653Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/host-device\"" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.957134346Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.957156034Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.957170108Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/host-local\"" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.969229928Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.969252711Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.969290414Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/ipvlan\"" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.977902389Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.977924733Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.977938935Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/loopback\"" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.985065680Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.985085861Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.985099423Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/macvlan\"" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.992479150Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.992498616Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.992512127Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/portmap\"" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.999875742Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:29 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.999894874Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:29.999911361Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/ptp\"" Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.007622030Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.007645780Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.007660128Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/sbr\"" Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.015159199Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.015214599Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.015229771Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/static\"" Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.022273748Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.022296990Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.022314574Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/tuning\"" Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.030874838Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.030988398Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.031072873Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/vlan\"" Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.038651724Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.038671302Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.038687282Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/vrf\"" Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.046249751Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.046282623Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.046299238Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_54e566b9-4267-42da-923a-4d7bcd088cef\"" Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:30.150313 2588 generic.go:296] "Generic (PLEG): container finished" podID=c0e47249-a5ff-460e-958d-53577387c154 containerID="db32d5f1c7edaafe4b12f6a178251fd11f7647122a044136c8d596c85418df5c" exitCode=0 Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:30.150378 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerDied Data:db32d5f1c7edaafe4b12f6a178251fd11f7647122a044136c8d596c85418df5c} Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.151245160Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7ad3b9824ea195f38b6346262f0c609415c487ffaeb554725621a8327266da0f" id=9f9f3db3-95e7-4186-98fe-19261952d329 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.151648092Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:4c7cce1e79d06b3bc147fc59bd0a72991be97e767193f661db48974f3bb9ac30,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7ad3b9824ea195f38b6346262f0c609415c487ffaeb554725621a8327266da0f],Size_:336119748,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=9f9f3db3-95e7-4186-98fe-19261952d329 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.152334727Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7ad3b9824ea195f38b6346262f0c609415c487ffaeb554725621a8327266da0f" id=a278c38f-95b0-43ed-aee2-b9b8c68101ec name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.152519997Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:4c7cce1e79d06b3bc147fc59bd0a72991be97e767193f661db48974f3bb9ac30,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7ad3b9824ea195f38b6346262f0c609415c487ffaeb554725621a8327266da0f],Size_:336119748,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=a278c38f-95b0-43ed-aee2-b9b8c68101ec name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.153469198Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/bond-cni-plugin" id=836133c2-5ff4-4377-b903-569550fdae44 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:30.153609 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-2r6nf" event=&{ID:aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f Type:ContainerStarted Data:be49f87e3a886d52fabc779de277793ff0cb222138bdd9809b5d437a60835b88} Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.153657512Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:30.153763 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-2r6nf" event=&{ID:aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f Type:ContainerStarted Data:53913239c5995dae15c3ae1ad17601117935102c6cb10665278cf79610336127} Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.496708883Z" level=info msg="Created container cf9d7ea85679d13ef09a1f895b6034dc9be7df433359fde59bb724fc05fd5d0a: openshift-multus/multus-additional-cni-plugins-5fgxc/bond-cni-plugin" id=836133c2-5ff4-4377-b903-569550fdae44 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.497165484Z" level=info msg="Starting container: cf9d7ea85679d13ef09a1f895b6034dc9be7df433359fde59bb724fc05fd5d0a" id=5164189a-ebb6-423d-bdaa-b677055e625c name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.504956884Z" level=info msg="Started container" PID=4343 containerID=cf9d7ea85679d13ef09a1f895b6034dc9be7df433359fde59bb724fc05fd5d0a description=openshift-multus/multus-additional-cni-plugins-5fgxc/bond-cni-plugin id=5164189a-ebb6-423d-bdaa-b677055e625c name=/runtime.v1.RuntimeService/StartContainer sandboxID=e9053289b0b05a28538d40135f5851bcbb4fef2e500409bd4a7c787c89dc9515 Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.525016171Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_bca7ec92-e193-4078-91a2-a42157266af2\"" Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.536768489Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.536790989Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.541403243Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/bond\"" Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.550771858Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.550799147Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:30 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:30.550817203Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_bca7ec92-e193-4078-91a2-a42157266af2\"" Jan 05 09:41:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:31.156392 2588 generic.go:296] "Generic (PLEG): container finished" podID=c0e47249-a5ff-460e-958d-53577387c154 containerID="cf9d7ea85679d13ef09a1f895b6034dc9be7df433359fde59bb724fc05fd5d0a" exitCode=0 Jan 05 09:41:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:31.156888 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerDied Data:cf9d7ea85679d13ef09a1f895b6034dc9be7df433359fde59bb724fc05fd5d0a} Jan 05 09:41:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:31.157838159Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:a8bcc8cf428e222c7e6677fe5f7f05e79da2aed1ded851cad696d83c94860b4c" id=d648117d-3713-48f4-a328-1116ab18c10e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:31.158066885Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:34fa51fa29c1ebd2a3ad620255bb70edf89b0a96348f08efac3dc2ca3200602d,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:a8bcc8cf428e222c7e6677fe5f7f05e79da2aed1ded851cad696d83c94860b4c],Size_:275275974,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=d648117d-3713-48f4-a328-1116ab18c10e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:31.158705926Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:a8bcc8cf428e222c7e6677fe5f7f05e79da2aed1ded851cad696d83c94860b4c" id=6789d936-3ae6-4e18-b1fc-2436448baa87 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:31.158870214Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:34fa51fa29c1ebd2a3ad620255bb70edf89b0a96348f08efac3dc2ca3200602d,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:a8bcc8cf428e222c7e6677fe5f7f05e79da2aed1ded851cad696d83c94860b4c],Size_:275275974,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=6789d936-3ae6-4e18-b1fc-2436448baa87 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:31.159595464Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/routeoverride-cni" id=3f381cd5-a99e-471b-b5b1-ddd097b15ddc name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:31.159718385Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:31.453378830Z" level=info msg="Created container c84024fb82e27d7f743b8d73bcdd8c77909d8a967fa5b3eda60d065878f668a1: openshift-multus/multus-additional-cni-plugins-5fgxc/routeoverride-cni" id=3f381cd5-a99e-471b-b5b1-ddd097b15ddc name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:31.454168095Z" level=info msg="Starting container: c84024fb82e27d7f743b8d73bcdd8c77909d8a967fa5b3eda60d065878f668a1" id=b67c060e-d9a7-488d-91a8-63a680e0ff87 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:41:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:31.467989836Z" level=info msg="Started container" PID=4408 containerID=c84024fb82e27d7f743b8d73bcdd8c77909d8a967fa5b3eda60d065878f668a1 description=openshift-multus/multus-additional-cni-plugins-5fgxc/routeoverride-cni id=b67c060e-d9a7-488d-91a8-63a680e0ff87 name=/runtime.v1.RuntimeService/StartContainer sandboxID=e9053289b0b05a28538d40135f5851bcbb4fef2e500409bd4a7c787c89dc9515 Jan 05 09:41:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:31.507471066Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_08f5f151-1cd3-4802-961e-30d68532685f\"" Jan 05 09:41:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:31.518109360Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:31.518141695Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:31.521056368Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/route-override\"" Jan 05 09:41:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:31.532403545Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:31.532432759Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:31.532449611Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_08f5f151-1cd3-4802-961e-30d68532685f\"" Jan 05 09:41:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:31.640595516Z" level=info msg="cleanup sandbox network" Jan 05 09:41:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:32.139824 2588 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-sdn/sdn-66nsp" Jan 05 09:41:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:32.159859 2588 generic.go:296] "Generic (PLEG): container finished" podID=c0e47249-a5ff-460e-958d-53577387c154 containerID="c84024fb82e27d7f743b8d73bcdd8c77909d8a967fa5b3eda60d065878f668a1" exitCode=0 Jan 05 09:41:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:32.159889 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerDied Data:c84024fb82e27d7f743b8d73bcdd8c77909d8a967fa5b3eda60d065878f668a1} Jan 05 09:41:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:32.160869906Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451" id=12ed9164-9ba1-46f2-af33-7cc72937699d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:32.161134294Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:56f4df3c126a569448a12d5112b7c20e99e219b2b2fe6a04c4ba743396fd7751,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451],Size_:445620510,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=12ed9164-9ba1-46f2-af33-7cc72937699d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:32.161797408Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451" id=c8a52e04-d72e-47ce-8088-c4d9f9fcc8a8 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:32.161983180Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:56f4df3c126a569448a12d5112b7c20e99e219b2b2fe6a04c4ba743396fd7751,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451],Size_:445620510,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=c8a52e04-d72e-47ce-8088-c4d9f9fcc8a8 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:32.162598126Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/whereabouts-cni-bincopy" id=9fb5d9c5-02c3-4562-a665-3a46539ba922 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:32.162719551Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:32.247854 2588 plugin_watcher.go:203] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/diskplugin.csi.alibabacloud.com-reg.sock" Jan 05 09:41:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:32.341822473Z" level=info msg="Created container 3f4061577652f73a57059601ec13dee108fe3ad1994234418e3b362d0bd8a29f: openshift-multus/multus-additional-cni-plugins-5fgxc/whereabouts-cni-bincopy" id=9fb5d9c5-02c3-4562-a665-3a46539ba922 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:32.342288649Z" level=info msg="Starting container: 3f4061577652f73a57059601ec13dee108fe3ad1994234418e3b362d0bd8a29f" id=505dad8a-c6ae-45de-ab02-7511cf721b5c name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:41:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:32.349338134Z" level=info msg="Started container" PID=4499 containerID=3f4061577652f73a57059601ec13dee108fe3ad1994234418e3b362d0bd8a29f description=openshift-multus/multus-additional-cni-plugins-5fgxc/whereabouts-cni-bincopy id=505dad8a-c6ae-45de-ab02-7511cf721b5c name=/runtime.v1.RuntimeService/StartContainer sandboxID=e9053289b0b05a28538d40135f5851bcbb4fef2e500409bd4a7c787c89dc9515 Jan 05 09:41:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:32.355916147Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_71615a14-a73b-4d0b-b926-2271b42c4e4f\"" Jan 05 09:41:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:32.368536288Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:32.368557758Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:32.451245145Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/whereabouts\"" Jan 05 09:41:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:32.461787344Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:41:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:32.461816626Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:41:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:32.461833489Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_71615a14-a73b-4d0b-b926-2271b42c4e4f\"" Jan 05 09:41:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:32.903817 2588 reconciler.go:164] "OperationExecutor.RegisterPlugin started" plugin={SocketPath:/var/lib/kubelet/plugins_registry/diskplugin.csi.alibabacloud.com-reg.sock Timestamp:2023-01-05 09:41:32.247875146 +0000 UTC m=+232.169708615 Handler: Name:} Jan 05 09:41:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:32.906880 2588 csi_plugin.go:99] kubernetes.io/csi: Trying to validate a new CSI Driver with name: diskplugin.csi.alibabacloud.com endpoint: /var/lib/kubelet/plugins/diskplugin.csi.alibabacloud.com/csi.sock versions: 1.0.0 Jan 05 09:41:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:32.906903 2588 csi_plugin.go:112] kubernetes.io/csi: Register new plugin with name: diskplugin.csi.alibabacloud.com at endpoint: /var/lib/kubelet/plugins/diskplugin.csi.alibabacloud.com/csi.sock Jan 05 09:41:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:33.165278 2588 generic.go:296] "Generic (PLEG): container finished" podID=c0e47249-a5ff-460e-958d-53577387c154 containerID="3f4061577652f73a57059601ec13dee108fe3ad1994234418e3b362d0bd8a29f" exitCode=0 Jan 05 09:41:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:33.165829 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerDied Data:3f4061577652f73a57059601ec13dee108fe3ad1994234418e3b362d0bd8a29f} Jan 05 09:41:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:33.166576109Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451" id=0a933fff-b8ab-4fe3-a12f-9037c528ac2e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:33.166823744Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:56f4df3c126a569448a12d5112b7c20e99e219b2b2fe6a04c4ba743396fd7751,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451],Size_:445620510,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=0a933fff-b8ab-4fe3-a12f-9037c528ac2e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:33.167565381Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451" id=4e4ef908-a307-4543-8d4d-26fc942d62e7 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:33.167768807Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:56f4df3c126a569448a12d5112b7c20e99e219b2b2fe6a04c4ba743396fd7751,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451],Size_:445620510,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=4e4ef908-a307-4543-8d4d-26fc942d62e7 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:33.168975494Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/whereabouts-cni" id=e0b2ab23-626e-4029-affe-36cf95575c68 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:33.169085221Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:33.427168488Z" level=info msg="Created container c7a310ebea8686cca731be40f4c60167c317f246378bb66043417e5d3df7c649: openshift-multus/multus-additional-cni-plugins-5fgxc/whereabouts-cni" id=e0b2ab23-626e-4029-affe-36cf95575c68 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:33.427727369Z" level=info msg="Starting container: c7a310ebea8686cca731be40f4c60167c317f246378bb66043417e5d3df7c649" id=5946a82c-3fc1-4c02-b057-9b09f8c012c9 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:41:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:33.436516405Z" level=info msg="Started container" PID=4643 containerID=c7a310ebea8686cca731be40f4c60167c317f246378bb66043417e5d3df7c649 description=openshift-multus/multus-additional-cni-plugins-5fgxc/whereabouts-cni id=5946a82c-3fc1-4c02-b057-9b09f8c012c9 name=/runtime.v1.RuntimeService/StartContainer sandboxID=e9053289b0b05a28538d40135f5851bcbb4fef2e500409bd4a7c787c89dc9515 Jan 05 09:41:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:34.169126 2588 generic.go:296] "Generic (PLEG): container finished" podID=c0e47249-a5ff-460e-958d-53577387c154 containerID="c7a310ebea8686cca731be40f4c60167c317f246378bb66043417e5d3df7c649" exitCode=0 Jan 05 09:41:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:34.169223 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerDied Data:c7a310ebea8686cca731be40f4c60167c317f246378bb66043417e5d3df7c649} Jan 05 09:41:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:34.170245878Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=6a949c0b-bb8d-497f-8276-9fae1d1cd416 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:34.170495533Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=6a949c0b-bb8d-497f-8276-9fae1d1cd416 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:34.171433539Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=41f47a93-88a7-42d0-bb55-624f508cdaa9 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:34.171605685Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=41f47a93-88a7-42d0-bb55-624f508cdaa9 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:41:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:34.172164387Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/kube-multus-additional-cni-plugins" id=8dc84cc9-9a79-4c7a-9aa8-241c337e2201 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:34.172315297Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:41:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:34.334295408Z" level=info msg="Created container d5fb53797503c324c9798f798dc512c161877f04f7be153de7f704b2b9975b35: openshift-multus/multus-additional-cni-plugins-5fgxc/kube-multus-additional-cni-plugins" id=8dc84cc9-9a79-4c7a-9aa8-241c337e2201 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:41:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:34.334854795Z" level=info msg="Starting container: d5fb53797503c324c9798f798dc512c161877f04f7be153de7f704b2b9975b35" id=82c94226-163a-4ca6-8282-caa2b408b3d8 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:41:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:34.342341184Z" level=info msg="Started container" PID=4831 containerID=d5fb53797503c324c9798f798dc512c161877f04f7be153de7f704b2b9975b35 description=openshift-multus/multus-additional-cni-plugins-5fgxc/kube-multus-additional-cni-plugins id=82c94226-163a-4ca6-8282-caa2b408b3d8 name=/runtime.v1.RuntimeService/StartContainer sandboxID=e9053289b0b05a28538d40135f5851bcbb4fef2e500409bd4a7c787c89dc9515 Jan 05 09:41:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:35.173302 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerStarted Data:d5fb53797503c324c9798f798dc512c161877f04f7be153de7f704b2b9975b35} Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.127719686Z" level=info msg="NetworkStart: stopping network for sandbox a4558df3b0fcc8b008b37af068d5302c4bced1a959f7e618f9a45abaabe07264" id=54140d90-7eb5-4850-b607-962c85c502b6 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.127985612Z" level=info msg="Got pod network &{Name:prometheus-operator-admission-webhook-598468bb49-94dsn Namespace:openshift-monitoring ID:a4558df3b0fcc8b008b37af068d5302c4bced1a959f7e618f9a45abaabe07264 UID:88745ab7-efe7-42c0-a2e8-75591518333d NetNS:/var/run/netns/90cf1dbd-7872-4b0c-9311-93ffbf2b4bf0 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.128038330Z" level=error msg="error loading cached network config: network \"multus-cni-network\" not found in CNI cache" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.128051780Z" level=warning msg="falling back to loading from existing plugins on disk" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.128064770Z" level=info msg="Deleting pod openshift-monitoring_prometheus-operator-admission-webhook-598468bb49-94dsn from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.137989872Z" level=info msg="NetworkStart: stopping network for sandbox 9994a437ea1715d6af20f45f2447fa709e3ded31fbff0b6e4cbdabe018ade0c1" id=b353c9f6-2798-4d08-8abf-2da0f599a503 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.138236991Z" level=info msg="Got pod network &{Name:network-metrics-daemon-hq594 Namespace:openshift-multus ID:9994a437ea1715d6af20f45f2447fa709e3ded31fbff0b6e4cbdabe018ade0c1 UID:50152296-9489-4fa8-aa42-7858debe1c08 NetNS:/var/run/netns/c543a3a8-aa1b-4690-abc6-2065756d32be Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.138293276Z" level=error msg="error loading cached network config: network \"multus-cni-network\" not found in CNI cache" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.138306667Z" level=warning msg="falling back to loading from existing plugins on disk" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.138319009Z" level=info msg="Deleting pod openshift-multus_network-metrics-daemon-hq594 from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.171981528Z" level=info msg="NetworkStart: stopping network for sandbox e386c433c8b6445e78f0a12c96e3583a65dc80ffd33df16f4209a33e828520f0" id=d77993bd-a749-4e24-b85d-15b672ca657a name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.172289445Z" level=info msg="Got pod network &{Name:telemeter-client-5598c645c4-mftdv Namespace:openshift-monitoring ID:e386c433c8b6445e78f0a12c96e3583a65dc80ffd33df16f4209a33e828520f0 UID:fc45e847-8f55-4c48-8c48-4bfee2419dca NetNS:/var/run/netns/aed17900-4e8c-44c2-8afd-9f1dab49d0cf Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.172335098Z" level=error msg="error loading cached network config: network \"multus-cni-network\" not found in CNI cache" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.172349430Z" level=warning msg="falling back to loading from existing plugins on disk" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.172362270Z" level=info msg="Deleting pod openshift-monitoring_telemeter-client-5598c645c4-mftdv from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:41:57Z [verbose] Del: openshift-monitoring:prometheus-operator-admission-webhook-598468bb49-94dsn:88745ab7-efe7-42c0-a2e8-75591518333d:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.195487689Z" level=info msg="runSandbox: cleaning up namespaces after failing to run sandbox a4558df3b0fcc8b008b37af068d5302c4bced1a959f7e618f9a45abaabe07264" id=54140d90-7eb5-4850-b607-962c85c502b6 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:41:57Z [verbose] Del: openshift-multus:network-metrics-daemon-hq594:50152296-9489-4fa8-aa42-7858debe1c08:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.209828431Z" level=info msg="runSandbox: cleaning up namespaces after failing to run sandbox 9994a437ea1715d6af20f45f2447fa709e3ded31fbff0b6e4cbdabe018ade0c1" id=b353c9f6-2798-4d08-8abf-2da0f599a503 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:41:57Z [verbose] Del: openshift-monitoring:telemeter-client-5598c645c4-mftdv:fc45e847-8f55-4c48-8c48-4bfee2419dca:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.237692760Z" level=info msg="runSandbox: cleaning up namespaces after failing to run sandbox e386c433c8b6445e78f0a12c96e3583a65dc80ffd33df16f4209a33e828520f0" id=d77993bd-a749-4e24-b85d-15b672ca657a name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.249349814Z" level=info msg="runSandbox: deleting pod ID a4558df3b0fcc8b008b37af068d5302c4bced1a959f7e618f9a45abaabe07264 from idIndex" id=54140d90-7eb5-4850-b607-962c85c502b6 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.249388537Z" level=info msg="runSandbox: removing pod sandbox a4558df3b0fcc8b008b37af068d5302c4bced1a959f7e618f9a45abaabe07264" id=54140d90-7eb5-4850-b607-962c85c502b6 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.249411486Z" level=info msg="runSandbox: deleting container ID from idIndex for sandbox a4558df3b0fcc8b008b37af068d5302c4bced1a959f7e618f9a45abaabe07264" id=54140d90-7eb5-4850-b607-962c85c502b6 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.249429106Z" level=info msg="runSandbox: unmounting shmPath for sandbox a4558df3b0fcc8b008b37af068d5302c4bced1a959f7e618f9a45abaabe07264" id=54140d90-7eb5-4850-b607-962c85c502b6 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.255305212Z" level=info msg="runSandbox: removing pod sandbox from storage: a4558df3b0fcc8b008b37af068d5302c4bced1a959f7e618f9a45abaabe07264" id=54140d90-7eb5-4850-b607-962c85c502b6 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.255331640Z" level=info msg="runSandbox: deleting pod ID 9994a437ea1715d6af20f45f2447fa709e3ded31fbff0b6e4cbdabe018ade0c1 from idIndex" id=b353c9f6-2798-4d08-8abf-2da0f599a503 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.255359387Z" level=info msg="runSandbox: removing pod sandbox 9994a437ea1715d6af20f45f2447fa709e3ded31fbff0b6e4cbdabe018ade0c1" id=b353c9f6-2798-4d08-8abf-2da0f599a503 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.255380875Z" level=info msg="runSandbox: deleting container ID from idIndex for sandbox 9994a437ea1715d6af20f45f2447fa709e3ded31fbff0b6e4cbdabe018ade0c1" id=b353c9f6-2798-4d08-8abf-2da0f599a503 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.255400579Z" level=info msg="runSandbox: unmounting shmPath for sandbox 9994a437ea1715d6af20f45f2447fa709e3ded31fbff0b6e4cbdabe018ade0c1" id=b353c9f6-2798-4d08-8abf-2da0f599a503 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.260307764Z" level=info msg="runSandbox: removing pod sandbox from storage: 9994a437ea1715d6af20f45f2447fa709e3ded31fbff0b6e4cbdabe018ade0c1" id=b353c9f6-2798-4d08-8abf-2da0f599a503 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.260687342Z" level=info msg="runSandbox: releasing container name: k8s_POD_prometheus-operator-admission-webhook-598468bb49-94dsn_openshift-monitoring_88745ab7-efe7-42c0-a2e8-75591518333d_0" id=54140d90-7eb5-4850-b607-962c85c502b6 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.260716926Z" level=info msg="runSandbox: releasing pod sandbox name: k8s_prometheus-operator-admission-webhook-598468bb49-94dsn_openshift-monitoring_88745ab7-efe7-42c0-a2e8-75591518333d_0" id=54140d90-7eb5-4850-b607-962c85c502b6 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.262357 2588 remote_runtime.go:222] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_prometheus-operator-admission-webhook-598468bb49-94dsn_openshift-monitoring_88745ab7-efe7-42c0-a2e8-75591518333d_0(a4558df3b0fcc8b008b37af068d5302c4bced1a959f7e618f9a45abaabe07264): error adding pod openshift-monitoring_prometheus-operator-admission-webhook-598468bb49-94dsn to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn/88745ab7-efe7-42c0-a2e8-75591518333d]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.262449 2588 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_prometheus-operator-admission-webhook-598468bb49-94dsn_openshift-monitoring_88745ab7-efe7-42c0-a2e8-75591518333d_0(a4558df3b0fcc8b008b37af068d5302c4bced1a959f7e618f9a45abaabe07264): error adding pod openshift-monitoring_prometheus-operator-admission-webhook-598468bb49-94dsn to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn/88745ab7-efe7-42c0-a2e8-75591518333d]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.262493 2588 kuberuntime_manager.go:772] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_prometheus-operator-admission-webhook-598468bb49-94dsn_openshift-monitoring_88745ab7-efe7-42c0-a2e8-75591518333d_0(a4558df3b0fcc8b008b37af068d5302c4bced1a959f7e618f9a45abaabe07264): error adding pod openshift-monitoring_prometheus-operator-admission-webhook-598468bb49-94dsn to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn/88745ab7-efe7-42c0-a2e8-75591518333d]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.262575 2588 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"prometheus-operator-admission-webhook-598468bb49-94dsn_openshift-monitoring(88745ab7-efe7-42c0-a2e8-75591518333d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"prometheus-operator-admission-webhook-598468bb49-94dsn_openshift-monitoring(88745ab7-efe7-42c0-a2e8-75591518333d)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_prometheus-operator-admission-webhook-598468bb49-94dsn_openshift-monitoring_88745ab7-efe7-42c0-a2e8-75591518333d_0(a4558df3b0fcc8b008b37af068d5302c4bced1a959f7e618f9a45abaabe07264): error adding pod openshift-monitoring_prometheus-operator-admission-webhook-598468bb49-94dsn to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus\\\" name=\\\"multus-cni-network\\\" failed (add): Multus: [openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn/88745ab7-efe7-42c0-a2e8-75591518333d]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\"" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" podUID=88745ab7-efe7-42c0-a2e8-75591518333d Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.267493414Z" level=info msg="runSandbox: releasing container name: k8s_POD_network-metrics-daemon-hq594_openshift-multus_50152296-9489-4fa8-aa42-7858debe1c08_0" id=b353c9f6-2798-4d08-8abf-2da0f599a503 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.267518566Z" level=info msg="runSandbox: releasing pod sandbox name: k8s_network-metrics-daemon-hq594_openshift-multus_50152296-9489-4fa8-aa42-7858debe1c08_0" id=b353c9f6-2798-4d08-8abf-2da0f599a503 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.267700 2588 remote_runtime.go:222] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_network-metrics-daemon-hq594_openshift-multus_50152296-9489-4fa8-aa42-7858debe1c08_0(9994a437ea1715d6af20f45f2447fa709e3ded31fbff0b6e4cbdabe018ade0c1): error adding pod openshift-multus_network-metrics-daemon-hq594 to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-multus/network-metrics-daemon-hq594/50152296-9489-4fa8-aa42-7858debe1c08]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.267755 2588 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_network-metrics-daemon-hq594_openshift-multus_50152296-9489-4fa8-aa42-7858debe1c08_0(9994a437ea1715d6af20f45f2447fa709e3ded31fbff0b6e4cbdabe018ade0c1): error adding pod openshift-multus_network-metrics-daemon-hq594 to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-multus/network-metrics-daemon-hq594/50152296-9489-4fa8-aa42-7858debe1c08]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.267794 2588 kuberuntime_manager.go:772] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_network-metrics-daemon-hq594_openshift-multus_50152296-9489-4fa8-aa42-7858debe1c08_0(9994a437ea1715d6af20f45f2447fa709e3ded31fbff0b6e4cbdabe018ade0c1): error adding pod openshift-multus_network-metrics-daemon-hq594 to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-multus/network-metrics-daemon-hq594/50152296-9489-4fa8-aa42-7858debe1c08]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.267872 2588 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"network-metrics-daemon-hq594_openshift-multus(50152296-9489-4fa8-aa42-7858debe1c08)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"network-metrics-daemon-hq594_openshift-multus(50152296-9489-4fa8-aa42-7858debe1c08)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_network-metrics-daemon-hq594_openshift-multus_50152296-9489-4fa8-aa42-7858debe1c08_0(9994a437ea1715d6af20f45f2447fa709e3ded31fbff0b6e4cbdabe018ade0c1): error adding pod openshift-multus_network-metrics-daemon-hq594 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus\\\" name=\\\"multus-cni-network\\\" failed (add): Multus: [openshift-multus/network-metrics-daemon-hq594/50152296-9489-4fa8-aa42-7858debe1c08]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\"" pod="openshift-multus/network-metrics-daemon-hq594" podUID=50152296-9489-4fa8-aa42-7858debe1c08 Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.275956477Z" level=info msg="NetworkStart: stopping network for sandbox f2fab4d64b6bfb23dbb4eccdef95ed2b1090988efca6d9d87c9936f1c4416a63" id=0442c4c3-31eb-4586-8d53-4bf70c6ead84 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.276214767Z" level=info msg="Got pod network &{Name:alertmanager-main-0 Namespace:openshift-monitoring ID:f2fab4d64b6bfb23dbb4eccdef95ed2b1090988efca6d9d87c9936f1c4416a63 UID:05c9a6ea-140a-49a2-bde5-f220b05aa252 NetNS:/var/run/netns/171a3b30-0a69-40cf-bf44-3484be0306ea Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.276251652Z" level=error msg="error loading cached network config: network \"multus-cni-network\" not found in CNI cache" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.276273584Z" level=warning msg="falling back to loading from existing plugins on disk" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.276285163Z" level=info msg="Deleting pod openshift-monitoring_alertmanager-main-0 from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.284311918Z" level=info msg="runSandbox: deleting pod ID e386c433c8b6445e78f0a12c96e3583a65dc80ffd33df16f4209a33e828520f0 from idIndex" id=d77993bd-a749-4e24-b85d-15b672ca657a name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.284348217Z" level=info msg="runSandbox: removing pod sandbox e386c433c8b6445e78f0a12c96e3583a65dc80ffd33df16f4209a33e828520f0" id=d77993bd-a749-4e24-b85d-15b672ca657a name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.284370777Z" level=info msg="runSandbox: deleting container ID from idIndex for sandbox e386c433c8b6445e78f0a12c96e3583a65dc80ffd33df16f4209a33e828520f0" id=d77993bd-a749-4e24-b85d-15b672ca657a name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.284391210Z" level=info msg="runSandbox: unmounting shmPath for sandbox e386c433c8b6445e78f0a12c96e3583a65dc80ffd33df16f4209a33e828520f0" id=d77993bd-a749-4e24-b85d-15b672ca657a name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.291379198Z" level=info msg="runSandbox: removing pod sandbox from storage: e386c433c8b6445e78f0a12c96e3583a65dc80ffd33df16f4209a33e828520f0" id=d77993bd-a749-4e24-b85d-15b672ca657a name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.300399107Z" level=info msg="runSandbox: releasing container name: k8s_POD_telemeter-client-5598c645c4-mftdv_openshift-monitoring_fc45e847-8f55-4c48-8c48-4bfee2419dca_0" id=d77993bd-a749-4e24-b85d-15b672ca657a name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.300425080Z" level=info msg="runSandbox: releasing pod sandbox name: k8s_telemeter-client-5598c645c4-mftdv_openshift-monitoring_fc45e847-8f55-4c48-8c48-4bfee2419dca_0" id=d77993bd-a749-4e24-b85d-15b672ca657a name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.300653 2588 remote_runtime.go:222] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_telemeter-client-5598c645c4-mftdv_openshift-monitoring_fc45e847-8f55-4c48-8c48-4bfee2419dca_0(e386c433c8b6445e78f0a12c96e3583a65dc80ffd33df16f4209a33e828520f0): error adding pod openshift-monitoring_telemeter-client-5598c645c4-mftdv to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/telemeter-client-5598c645c4-mftdv/fc45e847-8f55-4c48-8c48-4bfee2419dca]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.300722 2588 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_telemeter-client-5598c645c4-mftdv_openshift-monitoring_fc45e847-8f55-4c48-8c48-4bfee2419dca_0(e386c433c8b6445e78f0a12c96e3583a65dc80ffd33df16f4209a33e828520f0): error adding pod openshift-monitoring_telemeter-client-5598c645c4-mftdv to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/telemeter-client-5598c645c4-mftdv/fc45e847-8f55-4c48-8c48-4bfee2419dca]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.300759 2588 kuberuntime_manager.go:772] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_telemeter-client-5598c645c4-mftdv_openshift-monitoring_fc45e847-8f55-4c48-8c48-4bfee2419dca_0(e386c433c8b6445e78f0a12c96e3583a65dc80ffd33df16f4209a33e828520f0): error adding pod openshift-monitoring_telemeter-client-5598c645c4-mftdv to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/telemeter-client-5598c645c4-mftdv/fc45e847-8f55-4c48-8c48-4bfee2419dca]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.300846 2588 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"telemeter-client-5598c645c4-mftdv_openshift-monitoring(fc45e847-8f55-4c48-8c48-4bfee2419dca)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"telemeter-client-5598c645c4-mftdv_openshift-monitoring(fc45e847-8f55-4c48-8c48-4bfee2419dca)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_telemeter-client-5598c645c4-mftdv_openshift-monitoring_fc45e847-8f55-4c48-8c48-4bfee2419dca_0(e386c433c8b6445e78f0a12c96e3583a65dc80ffd33df16f4209a33e828520f0): error adding pod openshift-monitoring_telemeter-client-5598c645c4-mftdv to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus\\\" name=\\\"multus-cni-network\\\" failed (add): Multus: [openshift-monitoring/telemeter-client-5598c645c4-mftdv/fc45e847-8f55-4c48-8c48-4bfee2419dca]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\"" pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" podUID=fc45e847-8f55-4c48-8c48-4bfee2419dca Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:41:57Z [verbose] Del: openshift-monitoring:alertmanager-main-0:05c9a6ea-140a-49a2-bde5-f220b05aa252:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.328246841Z" level=info msg="runSandbox: cleaning up namespaces after failing to run sandbox f2fab4d64b6bfb23dbb4eccdef95ed2b1090988efca6d9d87c9936f1c4416a63" id=0442c4c3-31eb-4586-8d53-4bf70c6ead84 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.342048523Z" level=info msg="NetworkStart: stopping network for sandbox a68ba2c2702e0f26953b0481e42659c0abc11c4f57afaf50c0d6c186047f92d4" id=881e2e19-4071-4b0d-9e73-71cd5edabda9 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.342317069Z" level=info msg="Got pod network &{Name:prometheus-adapter-5894cc86c7-qhmbh Namespace:openshift-monitoring ID:a68ba2c2702e0f26953b0481e42659c0abc11c4f57afaf50c0d6c186047f92d4 UID:a40bf20d-67a4-45e7-b701-ca5a75376c1f NetNS:/var/run/netns/0ef21f34-ce89-4cc1-b1bf-60bb857132a2 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.342354062Z" level=error msg="error loading cached network config: network \"multus-cni-network\" not found in CNI cache" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.342365607Z" level=warning msg="falling back to loading from existing plugins on disk" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.342377256Z" level=info msg="Deleting pod openshift-monitoring_prometheus-adapter-5894cc86c7-qhmbh from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:41:57Z [verbose] Del: openshift-monitoring:prometheus-adapter-5894cc86c7-qhmbh:a40bf20d-67a4-45e7-b701-ca5a75376c1f:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.383964107Z" level=info msg="runSandbox: cleaning up namespaces after failing to run sandbox a68ba2c2702e0f26953b0481e42659c0abc11c4f57afaf50c0d6c186047f92d4" id=881e2e19-4071-4b0d-9e73-71cd5edabda9 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.385318646Z" level=info msg="runSandbox: deleting pod ID f2fab4d64b6bfb23dbb4eccdef95ed2b1090988efca6d9d87c9936f1c4416a63 from idIndex" id=0442c4c3-31eb-4586-8d53-4bf70c6ead84 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.385358076Z" level=info msg="runSandbox: removing pod sandbox f2fab4d64b6bfb23dbb4eccdef95ed2b1090988efca6d9d87c9936f1c4416a63" id=0442c4c3-31eb-4586-8d53-4bf70c6ead84 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.385383264Z" level=info msg="runSandbox: deleting container ID from idIndex for sandbox f2fab4d64b6bfb23dbb4eccdef95ed2b1090988efca6d9d87c9936f1c4416a63" id=0442c4c3-31eb-4586-8d53-4bf70c6ead84 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.385402642Z" level=info msg="runSandbox: unmounting shmPath for sandbox f2fab4d64b6bfb23dbb4eccdef95ed2b1090988efca6d9d87c9936f1c4416a63" id=0442c4c3-31eb-4586-8d53-4bf70c6ead84 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.388296014Z" level=info msg="runSandbox: removing pod sandbox from storage: f2fab4d64b6bfb23dbb4eccdef95ed2b1090988efca6d9d87c9936f1c4416a63" id=0442c4c3-31eb-4586-8d53-4bf70c6ead84 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.394415100Z" level=info msg="runSandbox: releasing container name: k8s_POD_alertmanager-main-0_openshift-monitoring_05c9a6ea-140a-49a2-bde5-f220b05aa252_0" id=0442c4c3-31eb-4586-8d53-4bf70c6ead84 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.394443527Z" level=info msg="runSandbox: releasing pod sandbox name: k8s_alertmanager-main-0_openshift-monitoring_05c9a6ea-140a-49a2-bde5-f220b05aa252_0" id=0442c4c3-31eb-4586-8d53-4bf70c6ead84 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.394648 2588 remote_runtime.go:222] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_alertmanager-main-0_openshift-monitoring_05c9a6ea-140a-49a2-bde5-f220b05aa252_0(f2fab4d64b6bfb23dbb4eccdef95ed2b1090988efca6d9d87c9936f1c4416a63): error adding pod openshift-monitoring_alertmanager-main-0 to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/alertmanager-main-0/05c9a6ea-140a-49a2-bde5-f220b05aa252]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.394718 2588 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_alertmanager-main-0_openshift-monitoring_05c9a6ea-140a-49a2-bde5-f220b05aa252_0(f2fab4d64b6bfb23dbb4eccdef95ed2b1090988efca6d9d87c9936f1c4416a63): error adding pod openshift-monitoring_alertmanager-main-0 to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/alertmanager-main-0/05c9a6ea-140a-49a2-bde5-f220b05aa252]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.394755 2588 kuberuntime_manager.go:772] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_alertmanager-main-0_openshift-monitoring_05c9a6ea-140a-49a2-bde5-f220b05aa252_0(f2fab4d64b6bfb23dbb4eccdef95ed2b1090988efca6d9d87c9936f1c4416a63): error adding pod openshift-monitoring_alertmanager-main-0 to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/alertmanager-main-0/05c9a6ea-140a-49a2-bde5-f220b05aa252]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.394837 2588 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"alertmanager-main-0_openshift-monitoring(05c9a6ea-140a-49a2-bde5-f220b05aa252)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"alertmanager-main-0_openshift-monitoring(05c9a6ea-140a-49a2-bde5-f220b05aa252)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_alertmanager-main-0_openshift-monitoring_05c9a6ea-140a-49a2-bde5-f220b05aa252_0(f2fab4d64b6bfb23dbb4eccdef95ed2b1090988efca6d9d87c9936f1c4416a63): error adding pod openshift-monitoring_alertmanager-main-0 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus\\\" name=\\\"multus-cni-network\\\" failed (add): Multus: [openshift-monitoring/alertmanager-main-0/05c9a6ea-140a-49a2-bde5-f220b05aa252]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\"" pod="openshift-monitoring/alertmanager-main-0" podUID=05c9a6ea-140a-49a2-bde5-f220b05aa252 Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.412965024Z" level=info msg="NetworkStart: stopping network for sandbox bc4aa1a421b0c7450185bcf26f1acfb0471356c226b5d1eb5fea1eb8c214d725" id=5c911e85-e00a-4fdc-a522-11a401d26bd1 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.413182167Z" level=info msg="Got pod network &{Name:ingress-canary-8xrbh Namespace:openshift-ingress-canary ID:bc4aa1a421b0c7450185bcf26f1acfb0471356c226b5d1eb5fea1eb8c214d725 UID:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748 NetNS:/var/run/netns/48587174-1261-4ae7-908c-76565765a732 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.413246772Z" level=error msg="error loading cached network config: network \"multus-cni-network\" not found in CNI cache" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.413258146Z" level=warning msg="falling back to loading from existing plugins on disk" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.413276526Z" level=info msg="Deleting pod openshift-ingress-canary_ingress-canary-8xrbh from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.430059477Z" level=info msg="runSandbox: deleting pod ID a68ba2c2702e0f26953b0481e42659c0abc11c4f57afaf50c0d6c186047f92d4 from idIndex" id=881e2e19-4071-4b0d-9e73-71cd5edabda9 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.430113732Z" level=info msg="runSandbox: removing pod sandbox a68ba2c2702e0f26953b0481e42659c0abc11c4f57afaf50c0d6c186047f92d4" id=881e2e19-4071-4b0d-9e73-71cd5edabda9 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.430143398Z" level=info msg="runSandbox: deleting container ID from idIndex for sandbox a68ba2c2702e0f26953b0481e42659c0abc11c4f57afaf50c0d6c186047f92d4" id=881e2e19-4071-4b0d-9e73-71cd5edabda9 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.430169408Z" level=info msg="runSandbox: unmounting shmPath for sandbox a68ba2c2702e0f26953b0481e42659c0abc11c4f57afaf50c0d6c186047f92d4" id=881e2e19-4071-4b0d-9e73-71cd5edabda9 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.433311669Z" level=info msg="runSandbox: removing pod sandbox from storage: a68ba2c2702e0f26953b0481e42659c0abc11c4f57afaf50c0d6c186047f92d4" id=881e2e19-4071-4b0d-9e73-71cd5edabda9 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.440510740Z" level=info msg="runSandbox: releasing container name: k8s_POD_prometheus-adapter-5894cc86c7-qhmbh_openshift-monitoring_a40bf20d-67a4-45e7-b701-ca5a75376c1f_0" id=881e2e19-4071-4b0d-9e73-71cd5edabda9 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.440537798Z" level=info msg="runSandbox: releasing pod sandbox name: k8s_prometheus-adapter-5894cc86c7-qhmbh_openshift-monitoring_a40bf20d-67a4-45e7-b701-ca5a75376c1f_0" id=881e2e19-4071-4b0d-9e73-71cd5edabda9 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.440751 2588 remote_runtime.go:222] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_prometheus-adapter-5894cc86c7-qhmbh_openshift-monitoring_a40bf20d-67a4-45e7-b701-ca5a75376c1f_0(a68ba2c2702e0f26953b0481e42659c0abc11c4f57afaf50c0d6c186047f92d4): error adding pod openshift-monitoring_prometheus-adapter-5894cc86c7-qhmbh to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh/a40bf20d-67a4-45e7-b701-ca5a75376c1f]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.440818 2588 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_prometheus-adapter-5894cc86c7-qhmbh_openshift-monitoring_a40bf20d-67a4-45e7-b701-ca5a75376c1f_0(a68ba2c2702e0f26953b0481e42659c0abc11c4f57afaf50c0d6c186047f92d4): error adding pod openshift-monitoring_prometheus-adapter-5894cc86c7-qhmbh to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh/a40bf20d-67a4-45e7-b701-ca5a75376c1f]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.440857 2588 kuberuntime_manager.go:772] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_prometheus-adapter-5894cc86c7-qhmbh_openshift-monitoring_a40bf20d-67a4-45e7-b701-ca5a75376c1f_0(a68ba2c2702e0f26953b0481e42659c0abc11c4f57afaf50c0d6c186047f92d4): error adding pod openshift-monitoring_prometheus-adapter-5894cc86c7-qhmbh to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh/a40bf20d-67a4-45e7-b701-ca5a75376c1f]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.440943 2588 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"prometheus-adapter-5894cc86c7-qhmbh_openshift-monitoring(a40bf20d-67a4-45e7-b701-ca5a75376c1f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"prometheus-adapter-5894cc86c7-qhmbh_openshift-monitoring(a40bf20d-67a4-45e7-b701-ca5a75376c1f)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_prometheus-adapter-5894cc86c7-qhmbh_openshift-monitoring_a40bf20d-67a4-45e7-b701-ca5a75376c1f_0(a68ba2c2702e0f26953b0481e42659c0abc11c4f57afaf50c0d6c186047f92d4): error adding pod openshift-monitoring_prometheus-adapter-5894cc86c7-qhmbh to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus\\\" name=\\\"multus-cni-network\\\" failed (add): Multus: [openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh/a40bf20d-67a4-45e7-b701-ca5a75376c1f]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\"" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" podUID=a40bf20d-67a4-45e7-b701-ca5a75376c1f Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.443120847Z" level=info msg="NetworkStart: stopping network for sandbox b9d09b59ecb8e2b530349c78a99778d6d8fc71efffc651d62a1e3fbb2005564d" id=c89c2627-bb06-4db8-b345-71d0fc5cb42c name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.443371675Z" level=info msg="Got pod network &{Name:thanos-querier-6f4f5fb7cf-rdm9l Namespace:openshift-monitoring ID:b9d09b59ecb8e2b530349c78a99778d6d8fc71efffc651d62a1e3fbb2005564d UID:7c63c9be-d1ad-44b6-8b33-09af3ee314af NetNS:/var/run/netns/7677b9cc-a641-408b-91a9-8419c8619347 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.443406634Z" level=error msg="error loading cached network config: network \"multus-cni-network\" not found in CNI cache" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.443418860Z" level=warning msg="falling back to loading from existing plugins on disk" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.443431086Z" level=info msg="Deleting pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:41:57Z [verbose] Del: openshift-ingress-canary:ingress-canary-8xrbh:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.487151779Z" level=info msg="runSandbox: cleaning up namespaces after failing to run sandbox bc4aa1a421b0c7450185bcf26f1acfb0471356c226b5d1eb5fea1eb8c214d725" id=5c911e85-e00a-4fdc-a522-11a401d26bd1 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.495399082Z" level=info msg="NetworkStart: stopping network for sandbox 56657e8828648ef03ee57a604c40642f3fe11d7eaf9907e27b69fd6d2ac4609e" id=c92ba848-1321-41e5-ab61-e6a211aaa3ab name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.495640114Z" level=info msg="Got pod network &{Name:kube-state-metrics-6798b548b9-xl88q Namespace:openshift-monitoring ID:56657e8828648ef03ee57a604c40642f3fe11d7eaf9907e27b69fd6d2ac4609e UID:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e NetNS:/var/run/netns/13a07410-2a86-4abc-b7cb-dddc3b44117e Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.495688079Z" level=error msg="error loading cached network config: network \"multus-cni-network\" not found in CNI cache" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.495704365Z" level=warning msg="falling back to loading from existing plugins on disk" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.495718920Z" level=info msg="Deleting pod openshift-monitoring_kube-state-metrics-6798b548b9-xl88q from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:41:57Z [verbose] Del: openshift-monitoring:thanos-querier-6f4f5fb7cf-rdm9l:7c63c9be-d1ad-44b6-8b33-09af3ee314af:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.505052977Z" level=info msg="runSandbox: cleaning up namespaces after failing to run sandbox b9d09b59ecb8e2b530349c78a99778d6d8fc71efffc651d62a1e3fbb2005564d" id=c89c2627-bb06-4db8-b345-71d0fc5cb42c name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.505063979Z" level=info msg="NetworkStart: stopping network for sandbox df8a985d595378c8d023c40bbf337dd18cbe474fb195d37d4cab28fb6e4e8e97" id=e708422a-b746-4e79-8e2e-c00d1dc48840 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.505407768Z" level=info msg="Got pod network &{Name:dns-default-2pfzf Namespace:openshift-dns ID:df8a985d595378c8d023c40bbf337dd18cbe474fb195d37d4cab28fb6e4e8e97 UID:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 NetNS:/var/run/netns/3bd81ca3-6cd1-42e1-b5af-cc2f9135e45d Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.505453938Z" level=error msg="error loading cached network config: network \"multus-cni-network\" not found in CNI cache" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.505466979Z" level=warning msg="falling back to loading from existing plugins on disk" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.505479337Z" level=info msg="Deleting pod openshift-dns_dns-default-2pfzf from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.524630817Z" level=info msg="runSandbox: deleting pod ID bc4aa1a421b0c7450185bcf26f1acfb0471356c226b5d1eb5fea1eb8c214d725 from idIndex" id=5c911e85-e00a-4fdc-a522-11a401d26bd1 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.524670149Z" level=info msg="runSandbox: removing pod sandbox bc4aa1a421b0c7450185bcf26f1acfb0471356c226b5d1eb5fea1eb8c214d725" id=5c911e85-e00a-4fdc-a522-11a401d26bd1 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.524692482Z" level=info msg="runSandbox: deleting container ID from idIndex for sandbox bc4aa1a421b0c7450185bcf26f1acfb0471356c226b5d1eb5fea1eb8c214d725" id=5c911e85-e00a-4fdc-a522-11a401d26bd1 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.524711737Z" level=info msg="runSandbox: unmounting shmPath for sandbox bc4aa1a421b0c7450185bcf26f1acfb0471356c226b5d1eb5fea1eb8c214d725" id=5c911e85-e00a-4fdc-a522-11a401d26bd1 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.525000123Z" level=info msg="NetworkStart: stopping network for sandbox f8841eb8a77afd8023660e94ffd508c50ea7025758849f56e786914d6fd4c8d1" id=fe43d99d-75da-4c33-84c4-0f91cfd3cfd4 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.525177893Z" level=info msg="Got pod network &{Name:openshift-state-metrics-6f5dcdf79b-wq256 Namespace:openshift-monitoring ID:f8841eb8a77afd8023660e94ffd508c50ea7025758849f56e786914d6fd4c8d1 UID:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 NetNS:/var/run/netns/82c20a6e-750c-4278-84a9-395a5a2ea52b Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.525231044Z" level=error msg="error loading cached network config: network \"multus-cni-network\" not found in CNI cache" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.525244086Z" level=warning msg="falling back to loading from existing plugins on disk" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.525254952Z" level=info msg="Deleting pod openshift-monitoring_openshift-state-metrics-6f5dcdf79b-wq256 from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.535451188Z" level=info msg="runSandbox: removing pod sandbox from storage: bc4aa1a421b0c7450185bcf26f1acfb0471356c226b5d1eb5fea1eb8c214d725" id=5c911e85-e00a-4fdc-a522-11a401d26bd1 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.537475030Z" level=info msg="NetworkStart: stopping network for sandbox 3307dc0da12fd52b662f10d53f8d87884254715041942b38a02e5c3023a51b72" id=d4ed3a90-2212-4ef9-8521-f3b9472cee6b name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.537737697Z" level=info msg="Got pod network &{Name:router-default-5cf6bbdff9-tnffs Namespace:openshift-ingress ID:3307dc0da12fd52b662f10d53f8d87884254715041942b38a02e5c3023a51b72 UID:c3b10ff4-a243-4a4e-b27d-f613e258666d NetNS:/var/run/netns/f01414fc-9125-434c-b723-d4b97fb30ecb Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.537779330Z" level=error msg="error loading cached network config: network \"multus-cni-network\" not found in CNI cache" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.537792436Z" level=warning msg="falling back to loading from existing plugins on disk" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.537804316Z" level=info msg="Deleting pod openshift-ingress_router-default-5cf6bbdff9-tnffs from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.571948016Z" level=info msg="runSandbox: releasing container name: k8s_POD_ingress-canary-8xrbh_openshift-ingress-canary_e04a47b4-aa8e-4d4e-9d45-aa8037dcb748_0" id=5c911e85-e00a-4fdc-a522-11a401d26bd1 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.572003027Z" level=info msg="runSandbox: releasing pod sandbox name: k8s_ingress-canary-8xrbh_openshift-ingress-canary_e04a47b4-aa8e-4d4e-9d45-aa8037dcb748_0" id=5c911e85-e00a-4fdc-a522-11a401d26bd1 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.573603 2588 remote_runtime.go:222] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_ingress-canary-8xrbh_openshift-ingress-canary_e04a47b4-aa8e-4d4e-9d45-aa8037dcb748_0(bc4aa1a421b0c7450185bcf26f1acfb0471356c226b5d1eb5fea1eb8c214d725): error adding pod openshift-ingress-canary_ingress-canary-8xrbh to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-ingress-canary/ingress-canary-8xrbh/e04a47b4-aa8e-4d4e-9d45-aa8037dcb748]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.573668 2588 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_ingress-canary-8xrbh_openshift-ingress-canary_e04a47b4-aa8e-4d4e-9d45-aa8037dcb748_0(bc4aa1a421b0c7450185bcf26f1acfb0471356c226b5d1eb5fea1eb8c214d725): error adding pod openshift-ingress-canary_ingress-canary-8xrbh to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-ingress-canary/ingress-canary-8xrbh/e04a47b4-aa8e-4d4e-9d45-aa8037dcb748]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-ingress-canary/ingress-canary-8xrbh" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.573709 2588 kuberuntime_manager.go:772] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_ingress-canary-8xrbh_openshift-ingress-canary_e04a47b4-aa8e-4d4e-9d45-aa8037dcb748_0(bc4aa1a421b0c7450185bcf26f1acfb0471356c226b5d1eb5fea1eb8c214d725): error adding pod openshift-ingress-canary_ingress-canary-8xrbh to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-ingress-canary/ingress-canary-8xrbh/e04a47b4-aa8e-4d4e-9d45-aa8037dcb748]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-ingress-canary/ingress-canary-8xrbh" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.573784 2588 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"ingress-canary-8xrbh_openshift-ingress-canary(e04a47b4-aa8e-4d4e-9d45-aa8037dcb748)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"ingress-canary-8xrbh_openshift-ingress-canary(e04a47b4-aa8e-4d4e-9d45-aa8037dcb748)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_ingress-canary-8xrbh_openshift-ingress-canary_e04a47b4-aa8e-4d4e-9d45-aa8037dcb748_0(bc4aa1a421b0c7450185bcf26f1acfb0471356c226b5d1eb5fea1eb8c214d725): error adding pod openshift-ingress-canary_ingress-canary-8xrbh to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus\\\" name=\\\"multus-cni-network\\\" failed (add): Multus: [openshift-ingress-canary/ingress-canary-8xrbh/e04a47b4-aa8e-4d4e-9d45-aa8037dcb748]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\"" pod="openshift-ingress-canary/ingress-canary-8xrbh" podUID=e04a47b4-aa8e-4d4e-9d45-aa8037dcb748 Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:41:57Z [verbose] Del: openshift-monitoring:kube-state-metrics-6798b548b9-xl88q:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.574349912Z" level=info msg="runSandbox: cleaning up namespaces after failing to run sandbox 56657e8828648ef03ee57a604c40642f3fe11d7eaf9907e27b69fd6d2ac4609e" id=c92ba848-1321-41e5-ab61-e6a211aaa3ab name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.574903725Z" level=info msg="runSandbox: deleting pod ID b9d09b59ecb8e2b530349c78a99778d6d8fc71efffc651d62a1e3fbb2005564d from idIndex" id=c89c2627-bb06-4db8-b345-71d0fc5cb42c name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.574957116Z" level=info msg="runSandbox: removing pod sandbox b9d09b59ecb8e2b530349c78a99778d6d8fc71efffc651d62a1e3fbb2005564d" id=c89c2627-bb06-4db8-b345-71d0fc5cb42c name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.574980709Z" level=info msg="runSandbox: deleting container ID from idIndex for sandbox b9d09b59ecb8e2b530349c78a99778d6d8fc71efffc651d62a1e3fbb2005564d" id=c89c2627-bb06-4db8-b345-71d0fc5cb42c name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.575001276Z" level=info msg="runSandbox: unmounting shmPath for sandbox b9d09b59ecb8e2b530349c78a99778d6d8fc71efffc651d62a1e3fbb2005564d" id=c89c2627-bb06-4db8-b345-71d0fc5cb42c name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.596400948Z" level=info msg="runSandbox: removing pod sandbox from storage: b9d09b59ecb8e2b530349c78a99778d6d8fc71efffc651d62a1e3fbb2005564d" id=c89c2627-bb06-4db8-b345-71d0fc5cb42c name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:41:57Z [verbose] Del: openshift-dns:dns-default-2pfzf:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.610177403Z" level=info msg="runSandbox: cleaning up namespaces after failing to run sandbox df8a985d595378c8d023c40bbf337dd18cbe474fb195d37d4cab28fb6e4e8e97" id=e708422a-b746-4e79-8e2e-c00d1dc48840 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.611618637Z" level=info msg="runSandbox: releasing container name: k8s_POD_thanos-querier-6f4f5fb7cf-rdm9l_openshift-monitoring_7c63c9be-d1ad-44b6-8b33-09af3ee314af_0" id=c89c2627-bb06-4db8-b345-71d0fc5cb42c name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.611652964Z" level=info msg="runSandbox: releasing pod sandbox name: k8s_thanos-querier-6f4f5fb7cf-rdm9l_openshift-monitoring_7c63c9be-d1ad-44b6-8b33-09af3ee314af_0" id=c89c2627-bb06-4db8-b345-71d0fc5cb42c name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.612308 2588 remote_runtime.go:222] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_thanos-querier-6f4f5fb7cf-rdm9l_openshift-monitoring_7c63c9be-d1ad-44b6-8b33-09af3ee314af_0(b9d09b59ecb8e2b530349c78a99778d6d8fc71efffc651d62a1e3fbb2005564d): error adding pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/7c63c9be-d1ad-44b6-8b33-09af3ee314af]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.612375 2588 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_thanos-querier-6f4f5fb7cf-rdm9l_openshift-monitoring_7c63c9be-d1ad-44b6-8b33-09af3ee314af_0(b9d09b59ecb8e2b530349c78a99778d6d8fc71efffc651d62a1e3fbb2005564d): error adding pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/7c63c9be-d1ad-44b6-8b33-09af3ee314af]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.612417 2588 kuberuntime_manager.go:772] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_thanos-querier-6f4f5fb7cf-rdm9l_openshift-monitoring_7c63c9be-d1ad-44b6-8b33-09af3ee314af_0(b9d09b59ecb8e2b530349c78a99778d6d8fc71efffc651d62a1e3fbb2005564d): error adding pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/7c63c9be-d1ad-44b6-8b33-09af3ee314af]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.612508 2588 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"thanos-querier-6f4f5fb7cf-rdm9l_openshift-monitoring(7c63c9be-d1ad-44b6-8b33-09af3ee314af)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"thanos-querier-6f4f5fb7cf-rdm9l_openshift-monitoring(7c63c9be-d1ad-44b6-8b33-09af3ee314af)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_thanos-querier-6f4f5fb7cf-rdm9l_openshift-monitoring_7c63c9be-d1ad-44b6-8b33-09af3ee314af_0(b9d09b59ecb8e2b530349c78a99778d6d8fc71efffc651d62a1e3fbb2005564d): error adding pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus\\\" name=\\\"multus-cni-network\\\" failed (add): Multus: [openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/7c63c9be-d1ad-44b6-8b33-09af3ee314af]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\"" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" podUID=7c63c9be-d1ad-44b6-8b33-09af3ee314af Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.617326793Z" level=info msg="NetworkStart: stopping network for sandbox 116a8242496b3ef676a5a564cd538412df8f6e3800ed923ea68d2687007a6d72" id=66686923-ed64-4f02-af98-ba9095064cea name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.617547732Z" level=info msg="Got pod network &{Name:network-check-target-xmq2g Namespace:openshift-network-diagnostics ID:116a8242496b3ef676a5a564cd538412df8f6e3800ed923ea68d2687007a6d72 UID:f364a949-09d7-4cfb-83ff-e532b822a557 NetNS:/var/run/netns/88a729ff-9dec-4eae-9c39-4d6236f2267e Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.617589936Z" level=error msg="error loading cached network config: network \"multus-cni-network\" not found in CNI cache" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.617602304Z" level=warning msg="falling back to loading from existing plugins on disk" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.617613423Z" level=info msg="Deleting pod openshift-network-diagnostics_network-check-target-xmq2g from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:41:57Z [verbose] Del: openshift-monitoring:openshift-state-metrics-6f5dcdf79b-wq256:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.626743685Z" level=info msg="runSandbox: cleaning up namespaces after failing to run sandbox f8841eb8a77afd8023660e94ffd508c50ea7025758849f56e786914d6fd4c8d1" id=fe43d99d-75da-4c33-84c4-0f91cfd3cfd4 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:41:57Z [verbose] Del: openshift-ingress:router-default-5cf6bbdff9-tnffs:c3b10ff4-a243-4a4e-b27d-f613e258666d:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.636283476Z" level=info msg="runSandbox: cleaning up namespaces after failing to run sandbox 3307dc0da12fd52b662f10d53f8d87884254715041942b38a02e5c3023a51b72" id=d4ed3a90-2212-4ef9-8521-f3b9472cee6b name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.645340837Z" level=info msg="runSandbox: deleting pod ID 56657e8828648ef03ee57a604c40642f3fe11d7eaf9907e27b69fd6d2ac4609e from idIndex" id=c92ba848-1321-41e5-ab61-e6a211aaa3ab name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.645379969Z" level=info msg="runSandbox: removing pod sandbox 56657e8828648ef03ee57a604c40642f3fe11d7eaf9907e27b69fd6d2ac4609e" id=c92ba848-1321-41e5-ab61-e6a211aaa3ab name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.645418582Z" level=info msg="runSandbox: deleting container ID from idIndex for sandbox 56657e8828648ef03ee57a604c40642f3fe11d7eaf9907e27b69fd6d2ac4609e" id=c92ba848-1321-41e5-ab61-e6a211aaa3ab name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.645440057Z" level=info msg="runSandbox: unmounting shmPath for sandbox 56657e8828648ef03ee57a604c40642f3fe11d7eaf9907e27b69fd6d2ac4609e" id=c92ba848-1321-41e5-ab61-e6a211aaa3ab name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.668938653Z" level=info msg="runSandbox: removing pod sandbox from storage: 56657e8828648ef03ee57a604c40642f3fe11d7eaf9907e27b69fd6d2ac4609e" id=c92ba848-1321-41e5-ab61-e6a211aaa3ab name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.675434051Z" level=info msg="runSandbox: deleting pod ID df8a985d595378c8d023c40bbf337dd18cbe474fb195d37d4cab28fb6e4e8e97 from idIndex" id=e708422a-b746-4e79-8e2e-c00d1dc48840 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.675476298Z" level=info msg="runSandbox: removing pod sandbox df8a985d595378c8d023c40bbf337dd18cbe474fb195d37d4cab28fb6e4e8e97" id=e708422a-b746-4e79-8e2e-c00d1dc48840 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.675500755Z" level=info msg="runSandbox: deleting container ID from idIndex for sandbox df8a985d595378c8d023c40bbf337dd18cbe474fb195d37d4cab28fb6e4e8e97" id=e708422a-b746-4e79-8e2e-c00d1dc48840 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.675522365Z" level=info msg="runSandbox: unmounting shmPath for sandbox df8a985d595378c8d023c40bbf337dd18cbe474fb195d37d4cab28fb6e4e8e97" id=e708422a-b746-4e79-8e2e-c00d1dc48840 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:41:57Z [verbose] Del: openshift-network-diagnostics:network-check-target-xmq2g:f364a949-09d7-4cfb-83ff-e532b822a557:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.677354384Z" level=info msg="runSandbox: cleaning up namespaces after failing to run sandbox 116a8242496b3ef676a5a564cd538412df8f6e3800ed923ea68d2687007a6d72" id=66686923-ed64-4f02-af98-ba9095064cea name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.677581400Z" level=info msg="runSandbox: removing pod sandbox from storage: df8a985d595378c8d023c40bbf337dd18cbe474fb195d37d4cab28fb6e4e8e97" id=e708422a-b746-4e79-8e2e-c00d1dc48840 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.678529333Z" level=info msg="runSandbox: releasing container name: k8s_POD_kube-state-metrics-6798b548b9-xl88q_openshift-monitoring_2ecd2e2c-ab52-42e0-8fde-64e45804cc9e_0" id=c92ba848-1321-41e5-ab61-e6a211aaa3ab name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.678553252Z" level=info msg="runSandbox: releasing pod sandbox name: k8s_kube-state-metrics-6798b548b9-xl88q_openshift-monitoring_2ecd2e2c-ab52-42e0-8fde-64e45804cc9e_0" id=c92ba848-1321-41e5-ab61-e6a211aaa3ab name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.680084 2588 remote_runtime.go:222] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-state-metrics-6798b548b9-xl88q_openshift-monitoring_2ecd2e2c-ab52-42e0-8fde-64e45804cc9e_0(56657e8828648ef03ee57a604c40642f3fe11d7eaf9907e27b69fd6d2ac4609e): error adding pod openshift-monitoring_kube-state-metrics-6798b548b9-xl88q to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.680163 2588 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-state-metrics-6798b548b9-xl88q_openshift-monitoring_2ecd2e2c-ab52-42e0-8fde-64e45804cc9e_0(56657e8828648ef03ee57a604c40642f3fe11d7eaf9907e27b69fd6d2ac4609e): error adding pod openshift-monitoring_kube-state-metrics-6798b548b9-xl88q to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.680221 2588 kuberuntime_manager.go:772] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-state-metrics-6798b548b9-xl88q_openshift-monitoring_2ecd2e2c-ab52-42e0-8fde-64e45804cc9e_0(56657e8828648ef03ee57a604c40642f3fe11d7eaf9907e27b69fd6d2ac4609e): error adding pod openshift-monitoring_kube-state-metrics-6798b548b9-xl88q to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.680335 2588 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"kube-state-metrics-6798b548b9-xl88q_openshift-monitoring(2ecd2e2c-ab52-42e0-8fde-64e45804cc9e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"kube-state-metrics-6798b548b9-xl88q_openshift-monitoring(2ecd2e2c-ab52-42e0-8fde-64e45804cc9e)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-state-metrics-6798b548b9-xl88q_openshift-monitoring_2ecd2e2c-ab52-42e0-8fde-64e45804cc9e_0(56657e8828648ef03ee57a604c40642f3fe11d7eaf9907e27b69fd6d2ac4609e): error adding pod openshift-monitoring_kube-state-metrics-6798b548b9-xl88q to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus\\\" name=\\\"multus-cni-network\\\" failed (add): Multus: [openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\"" pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" podUID=2ecd2e2c-ab52-42e0-8fde-64e45804cc9e Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.683310754Z" level=info msg="runSandbox: deleting pod ID f8841eb8a77afd8023660e94ffd508c50ea7025758849f56e786914d6fd4c8d1 from idIndex" id=fe43d99d-75da-4c33-84c4-0f91cfd3cfd4 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.683343826Z" level=info msg="runSandbox: removing pod sandbox f8841eb8a77afd8023660e94ffd508c50ea7025758849f56e786914d6fd4c8d1" id=fe43d99d-75da-4c33-84c4-0f91cfd3cfd4 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.683365004Z" level=info msg="runSandbox: deleting container ID from idIndex for sandbox f8841eb8a77afd8023660e94ffd508c50ea7025758849f56e786914d6fd4c8d1" id=fe43d99d-75da-4c33-84c4-0f91cfd3cfd4 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.683385238Z" level=info msg="runSandbox: unmounting shmPath for sandbox f8841eb8a77afd8023660e94ffd508c50ea7025758849f56e786914d6fd4c8d1" id=fe43d99d-75da-4c33-84c4-0f91cfd3cfd4 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.685581960Z" level=info msg="runSandbox: removing pod sandbox from storage: f8841eb8a77afd8023660e94ffd508c50ea7025758849f56e786914d6fd4c8d1" id=fe43d99d-75da-4c33-84c4-0f91cfd3cfd4 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.685591002Z" level=info msg="runSandbox: releasing container name: k8s_POD_dns-default-2pfzf_openshift-dns_1cc2aadb-14f2-4d6e-97fc-6fdab8889a63_0" id=e708422a-b746-4e79-8e2e-c00d1dc48840 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.685743177Z" level=info msg="runSandbox: releasing pod sandbox name: k8s_dns-default-2pfzf_openshift-dns_1cc2aadb-14f2-4d6e-97fc-6fdab8889a63_0" id=e708422a-b746-4e79-8e2e-c00d1dc48840 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.687035749Z" level=info msg="runSandbox: deleting pod ID 3307dc0da12fd52b662f10d53f8d87884254715041942b38a02e5c3023a51b72 from idIndex" id=d4ed3a90-2212-4ef9-8521-f3b9472cee6b name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.687070328Z" level=info msg="runSandbox: removing pod sandbox 3307dc0da12fd52b662f10d53f8d87884254715041942b38a02e5c3023a51b72" id=d4ed3a90-2212-4ef9-8521-f3b9472cee6b name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.687091789Z" level=info msg="runSandbox: deleting container ID from idIndex for sandbox 3307dc0da12fd52b662f10d53f8d87884254715041942b38a02e5c3023a51b72" id=d4ed3a90-2212-4ef9-8521-f3b9472cee6b name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.687111740Z" level=info msg="runSandbox: unmounting shmPath for sandbox 3307dc0da12fd52b662f10d53f8d87884254715041942b38a02e5c3023a51b72" id=d4ed3a90-2212-4ef9-8521-f3b9472cee6b name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.687623 2588 remote_runtime.go:222] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_dns-default-2pfzf_openshift-dns_1cc2aadb-14f2-4d6e-97fc-6fdab8889a63_0(df8a985d595378c8d023c40bbf337dd18cbe474fb195d37d4cab28fb6e4e8e97): error adding pod openshift-dns_dns-default-2pfzf to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-dns/dns-default-2pfzf/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.687693 2588 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_dns-default-2pfzf_openshift-dns_1cc2aadb-14f2-4d6e-97fc-6fdab8889a63_0(df8a985d595378c8d023c40bbf337dd18cbe474fb195d37d4cab28fb6e4e8e97): error adding pod openshift-dns_dns-default-2pfzf to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-dns/dns-default-2pfzf/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-dns/dns-default-2pfzf" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.687728 2588 kuberuntime_manager.go:772] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_dns-default-2pfzf_openshift-dns_1cc2aadb-14f2-4d6e-97fc-6fdab8889a63_0(df8a985d595378c8d023c40bbf337dd18cbe474fb195d37d4cab28fb6e4e8e97): error adding pod openshift-dns_dns-default-2pfzf to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-dns/dns-default-2pfzf/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-dns/dns-default-2pfzf" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.687805 2588 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"dns-default-2pfzf_openshift-dns(1cc2aadb-14f2-4d6e-97fc-6fdab8889a63)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"dns-default-2pfzf_openshift-dns(1cc2aadb-14f2-4d6e-97fc-6fdab8889a63)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_dns-default-2pfzf_openshift-dns_1cc2aadb-14f2-4d6e-97fc-6fdab8889a63_0(df8a985d595378c8d023c40bbf337dd18cbe474fb195d37d4cab28fb6e4e8e97): error adding pod openshift-dns_dns-default-2pfzf to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus\\\" name=\\\"multus-cni-network\\\" failed (add): Multus: [openshift-dns/dns-default-2pfzf/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\"" pod="openshift-dns/dns-default-2pfzf" podUID=1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.691336332Z" level=info msg="runSandbox: removing pod sandbox from storage: 3307dc0da12fd52b662f10d53f8d87884254715041942b38a02e5c3023a51b72" id=d4ed3a90-2212-4ef9-8521-f3b9472cee6b name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.691485020Z" level=info msg="runSandbox: releasing container name: k8s_POD_openshift-state-metrics-6f5dcdf79b-wq256_openshift-monitoring_9de5de36-9b80-4a0b-a615-f0f9ccda2ff6_0" id=fe43d99d-75da-4c33-84c4-0f91cfd3cfd4 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.691509408Z" level=info msg="runSandbox: releasing pod sandbox name: k8s_openshift-state-metrics-6f5dcdf79b-wq256_openshift-monitoring_9de5de36-9b80-4a0b-a615-f0f9ccda2ff6_0" id=fe43d99d-75da-4c33-84c4-0f91cfd3cfd4 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.691720 2588 remote_runtime.go:222] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-state-metrics-6f5dcdf79b-wq256_openshift-monitoring_9de5de36-9b80-4a0b-a615-f0f9ccda2ff6_0(f8841eb8a77afd8023660e94ffd508c50ea7025758849f56e786914d6fd4c8d1): error adding pod openshift-monitoring_openshift-state-metrics-6f5dcdf79b-wq256 to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.691775 2588 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-state-metrics-6f5dcdf79b-wq256_openshift-monitoring_9de5de36-9b80-4a0b-a615-f0f9ccda2ff6_0(f8841eb8a77afd8023660e94ffd508c50ea7025758849f56e786914d6fd4c8d1): error adding pod openshift-monitoring_openshift-state-metrics-6f5dcdf79b-wq256 to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.691815 2588 kuberuntime_manager.go:772] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-state-metrics-6f5dcdf79b-wq256_openshift-monitoring_9de5de36-9b80-4a0b-a615-f0f9ccda2ff6_0(f8841eb8a77afd8023660e94ffd508c50ea7025758849f56e786914d6fd4c8d1): error adding pod openshift-monitoring_openshift-state-metrics-6f5dcdf79b-wq256 to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.691893 2588 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"openshift-state-metrics-6f5dcdf79b-wq256_openshift-monitoring(9de5de36-9b80-4a0b-a615-f0f9ccda2ff6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"openshift-state-metrics-6f5dcdf79b-wq256_openshift-monitoring(9de5de36-9b80-4a0b-a615-f0f9ccda2ff6)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-state-metrics-6f5dcdf79b-wq256_openshift-monitoring_9de5de36-9b80-4a0b-a615-f0f9ccda2ff6_0(f8841eb8a77afd8023660e94ffd508c50ea7025758849f56e786914d6fd4c8d1): error adding pod openshift-monitoring_openshift-state-metrics-6f5dcdf79b-wq256 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus\\\" name=\\\"multus-cni-network\\\" failed (add): Multus: [openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\"" pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" podUID=9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.717950513Z" level=info msg="runSandbox: releasing container name: k8s_POD_router-default-5cf6bbdff9-tnffs_openshift-ingress_c3b10ff4-a243-4a4e-b27d-f613e258666d_0" id=d4ed3a90-2212-4ef9-8521-f3b9472cee6b name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.718006772Z" level=info msg="runSandbox: releasing pod sandbox name: k8s_router-default-5cf6bbdff9-tnffs_openshift-ingress_c3b10ff4-a243-4a4e-b27d-f613e258666d_0" id=d4ed3a90-2212-4ef9-8521-f3b9472cee6b name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.726409 2588 remote_runtime.go:222] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_router-default-5cf6bbdff9-tnffs_openshift-ingress_c3b10ff4-a243-4a4e-b27d-f613e258666d_0(3307dc0da12fd52b662f10d53f8d87884254715041942b38a02e5c3023a51b72): error adding pod openshift-ingress_router-default-5cf6bbdff9-tnffs to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-ingress/router-default-5cf6bbdff9-tnffs/c3b10ff4-a243-4a4e-b27d-f613e258666d]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.726477 2588 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_router-default-5cf6bbdff9-tnffs_openshift-ingress_c3b10ff4-a243-4a4e-b27d-f613e258666d_0(3307dc0da12fd52b662f10d53f8d87884254715041942b38a02e5c3023a51b72): error adding pod openshift-ingress_router-default-5cf6bbdff9-tnffs to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-ingress/router-default-5cf6bbdff9-tnffs/c3b10ff4-a243-4a4e-b27d-f613e258666d]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.726519 2588 kuberuntime_manager.go:772] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_router-default-5cf6bbdff9-tnffs_openshift-ingress_c3b10ff4-a243-4a4e-b27d-f613e258666d_0(3307dc0da12fd52b662f10d53f8d87884254715041942b38a02e5c3023a51b72): error adding pod openshift-ingress_router-default-5cf6bbdff9-tnffs to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-ingress/router-default-5cf6bbdff9-tnffs/c3b10ff4-a243-4a4e-b27d-f613e258666d]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.726595 2588 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"router-default-5cf6bbdff9-tnffs_openshift-ingress(c3b10ff4-a243-4a4e-b27d-f613e258666d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"router-default-5cf6bbdff9-tnffs_openshift-ingress(c3b10ff4-a243-4a4e-b27d-f613e258666d)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_router-default-5cf6bbdff9-tnffs_openshift-ingress_c3b10ff4-a243-4a4e-b27d-f613e258666d_0(3307dc0da12fd52b662f10d53f8d87884254715041942b38a02e5c3023a51b72): error adding pod openshift-ingress_router-default-5cf6bbdff9-tnffs to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus\\\" name=\\\"multus-cni-network\\\" failed (add): Multus: [openshift-ingress/router-default-5cf6bbdff9-tnffs/c3b10ff4-a243-4a4e-b27d-f613e258666d]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\"" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" podUID=c3b10ff4-a243-4a4e-b27d-f613e258666d Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.733313396Z" level=info msg="runSandbox: deleting pod ID 116a8242496b3ef676a5a564cd538412df8f6e3800ed923ea68d2687007a6d72 from idIndex" id=66686923-ed64-4f02-af98-ba9095064cea name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.733361219Z" level=info msg="runSandbox: removing pod sandbox 116a8242496b3ef676a5a564cd538412df8f6e3800ed923ea68d2687007a6d72" id=66686923-ed64-4f02-af98-ba9095064cea name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.733383868Z" level=info msg="runSandbox: deleting container ID from idIndex for sandbox 116a8242496b3ef676a5a564cd538412df8f6e3800ed923ea68d2687007a6d72" id=66686923-ed64-4f02-af98-ba9095064cea name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.733402335Z" level=info msg="runSandbox: unmounting shmPath for sandbox 116a8242496b3ef676a5a564cd538412df8f6e3800ed923ea68d2687007a6d72" id=66686923-ed64-4f02-af98-ba9095064cea name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.736287348Z" level=info msg="runSandbox: removing pod sandbox from storage: 116a8242496b3ef676a5a564cd538412df8f6e3800ed923ea68d2687007a6d72" id=66686923-ed64-4f02-af98-ba9095064cea name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.742534134Z" level=info msg="runSandbox: releasing container name: k8s_POD_network-check-target-xmq2g_openshift-network-diagnostics_f364a949-09d7-4cfb-83ff-e532b822a557_0" id=66686923-ed64-4f02-af98-ba9095064cea name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:41:57.742561278Z" level=info msg="runSandbox: releasing pod sandbox name: k8s_network-check-target-xmq2g_openshift-network-diagnostics_f364a949-09d7-4cfb-83ff-e532b822a557_0" id=66686923-ed64-4f02-af98-ba9095064cea name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.742779 2588 remote_runtime.go:222] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_network-check-target-xmq2g_openshift-network-diagnostics_f364a949-09d7-4cfb-83ff-e532b822a557_0(116a8242496b3ef676a5a564cd538412df8f6e3800ed923ea68d2687007a6d72): error adding pod openshift-network-diagnostics_network-check-target-xmq2g to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-network-diagnostics/network-check-target-xmq2g/f364a949-09d7-4cfb-83ff-e532b822a557]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.742840 2588 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_network-check-target-xmq2g_openshift-network-diagnostics_f364a949-09d7-4cfb-83ff-e532b822a557_0(116a8242496b3ef676a5a564cd538412df8f6e3800ed923ea68d2687007a6d72): error adding pod openshift-network-diagnostics_network-check-target-xmq2g to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-network-diagnostics/network-check-target-xmq2g/f364a949-09d7-4cfb-83ff-e532b822a557]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.742876 2588 kuberuntime_manager.go:772] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_network-check-target-xmq2g_openshift-network-diagnostics_f364a949-09d7-4cfb-83ff-e532b822a557_0(116a8242496b3ef676a5a564cd538412df8f6e3800ed923ea68d2687007a6d72): error adding pod openshift-network-diagnostics_network-check-target-xmq2g to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-network-diagnostics/network-check-target-xmq2g/f364a949-09d7-4cfb-83ff-e532b822a557]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 09:41:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:41:57.742971 2588 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"network-check-target-xmq2g_openshift-network-diagnostics(f364a949-09d7-4cfb-83ff-e532b822a557)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"network-check-target-xmq2g_openshift-network-diagnostics(f364a949-09d7-4cfb-83ff-e532b822a557)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_network-check-target-xmq2g_openshift-network-diagnostics_f364a949-09d7-4cfb-83ff-e532b822a557_0(116a8242496b3ef676a5a564cd538412df8f6e3800ed923ea68d2687007a6d72): error adding pod openshift-network-diagnostics_network-check-target-xmq2g to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus\\\" name=\\\"multus-cni-network\\\" failed (add): Multus: [openshift-network-diagnostics/network-check-target-xmq2g/f364a949-09d7-4cfb-83ff-e532b822a557]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\"" pod="openshift-network-diagnostics/network-check-target-xmq2g" podUID=f364a949-09d7-4cfb-83ff-e532b822a557 Jan 05 09:41:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:59.170138 2588 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-sdn_sdn-66nsp_d5b0f5e6-3d70-4315-b9c5-1403036f517b/sdn/1.log" Jan 05 09:41:59 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:41:59.206559 2588 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-sdn_sdn-66nsp_d5b0f5e6-3d70-4315-b9c5-1403036f517b/kube-rbac-proxy/1.log" Jan 05 09:42:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:08.517816 2588 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-dns_node-resolver-nxzr8_8f88cbe5-1221-4e6f-a6c9-f22da152b43f/dns-node-resolver/3.log" Jan 05 09:42:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:08.793807 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 09:42:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:08.794398 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" Jan 05 09:42:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:08.794857 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-8xrbh" Jan 05 09:42:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:08.795322 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:42:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:08.796100 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:42:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:08.796283855Z" level=info msg="Running pod sandbox: openshift-multus/network-metrics-daemon-hq594/POD" id=bdfb661b-c8b0-4ca4-88ff-e93c5b94ae90 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:08.796343488Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:08.796440666Z" level=info msg="Running pod sandbox: openshift-monitoring/alertmanager-main-0/POD" id=a553a87e-65f2-44c7-9aec-483691caf6f0 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:08.796474043Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:08.796564654Z" level=info msg="Running pod sandbox: openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn/POD" id=08d00f52-7155-4997-9c76-ed8b4000bc50 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:08.796591287Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:08.796928943Z" level=info msg="Running pod sandbox: openshift-ingress-canary/ingress-canary-8xrbh/POD" id=f2cc8080-3b7a-4eec-9f06-f162f82a9cba name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:08.796961274Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:08.797356647Z" level=info msg="Running pod sandbox: openshift-ingress/router-default-5cf6bbdff9-tnffs/POD" id=4fd824c4-281f-4b66-8101-8678d78afacb name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:08.797394413Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:08.833316956Z" level=info msg="Got pod network &{Name:alertmanager-main-0 Namespace:openshift-monitoring ID:d6d0c82bc9774e51559f70121d26a57c7c38cce6fb83f23b63a8bb671c408015 UID:05c9a6ea-140a-49a2-bde5-f220b05aa252 NetNS:/var/run/netns/c7a80f78-1128-4e9e-9931-f7ed057df8ad Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:42:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:08.833352993Z" level=info msg="Adding pod openshift-monitoring_alertmanager-main-0 to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:42:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:08.853630869Z" level=info msg="Got pod network &{Name:network-metrics-daemon-hq594 Namespace:openshift-multus ID:44ef9ac88935fc68d10995ab1204d3f92afbedeae9fabaf8c0abb737181fa9fd UID:50152296-9489-4fa8-aa42-7858debe1c08 NetNS:/var/run/netns/5a7d8972-1e35-45cd-add2-715bbc661172 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:42:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:08.853665719Z" level=info msg="Adding pod openshift-multus_network-metrics-daemon-hq594 to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:42:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:08.920293772Z" level=info msg="Got pod network &{Name:prometheus-operator-admission-webhook-598468bb49-94dsn Namespace:openshift-monitoring ID:e10e02d14faf7624d78743e0d1899ddf64963db130ecf653b0849bcde995a72c UID:88745ab7-efe7-42c0-a2e8-75591518333d NetNS:/var/run/netns/67866f67-3eb9-4569-aa7e-69d12f039663 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:42:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:08.920325110Z" level=info msg="Adding pod openshift-monitoring_prometheus-operator-admission-webhook-598468bb49-94dsn to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:42:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:08.943561341Z" level=info msg="Got pod network &{Name:router-default-5cf6bbdff9-tnffs Namespace:openshift-ingress ID:4814a25d28532df204d5f61d2aaabc14a67b24096511e9cd393f31ea5787ca12 UID:c3b10ff4-a243-4a4e-b27d-f613e258666d NetNS:/var/run/netns/d85fb3f0-7148-4415-ac9c-470bac5d14d4 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:42:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:08.943600347Z" level=info msg="Adding pod openshift-ingress_router-default-5cf6bbdff9-tnffs to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:42:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:08.995093869Z" level=info msg="Got pod network &{Name:ingress-canary-8xrbh Namespace:openshift-ingress-canary ID:2f42fdd8caa6c772e33c09307d855f2c6b126d16a962580bfad9af47fa893ef2 UID:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748 NetNS:/var/run/netns/db10749b-0af7-4faa-bbc1-64769e589a18 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:42:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:08.995143473Z" level=info msg="Adding pod openshift-ingress-canary_ingress-canary-8xrbh to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:42:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:09.793926 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:42:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:09.794483384Z" level=info msg="Running pod sandbox: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/POD" id=d5df9f78-035a-4776-a598-68baf0ff7872 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:09.794544675Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:09.813461615Z" level=info msg="Got pod network &{Name:thanos-querier-6f4f5fb7cf-rdm9l Namespace:openshift-monitoring ID:fcdbc115e8986f7ba4e1fc9bfa0728fd149a52733e5f0956b91d0b8262b33ca6 UID:7c63c9be-d1ad-44b6-8b33-09af3ee314af NetNS:/var/run/netns/d054aa63-f5ac-4893-88ff-2a7d67c9ea25 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:42:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:09.813490308Z" level=info msg="Adding pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:10.302739 2588 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-ingress/router-default-5cf6bbdff9-tnffs] Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:42:10Z [verbose] Add: openshift-ingress:router-default-5cf6bbdff9-tnffs:c3b10ff4-a243-4a4e-b27d-f613e258666d:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/d85fb3f0-7148-4415-ac9c-470bac5d14d4"}],"ips":[{"version":"4","interface":0,"address":"10.131.0.3/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: I0105 09:42:10.257886 5919 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-ingress", Name:"router-default-5cf6bbdff9-tnffs", UID:"c3b10ff4-a243-4a4e-b27d-f613e258666d", APIVersion:"v1", ResourceVersion:"96792", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.3/23] from openshift-sdn Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.309022480Z" level=info msg="Got pod network &{Name:router-default-5cf6bbdff9-tnffs Namespace:openshift-ingress ID:4814a25d28532df204d5f61d2aaabc14a67b24096511e9cd393f31ea5787ca12 UID:c3b10ff4-a243-4a4e-b27d-f613e258666d NetNS:/var/run/netns/d85fb3f0-7148-4415-ac9c-470bac5d14d4 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.309215120Z" level=info msg="Checking pod openshift-ingress_router-default-5cf6bbdff9-tnffs for CNI network multus-cni-network (type=multus)" Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:42:10.334973 2588 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3b10ff4_a243_4a4e_b27d_f613e258666d.slice/crio-4814a25d28532df204d5f61d2aaabc14a67b24096511e9cd393f31ea5787ca12.scope WatchSource:0}: Error finding container 4814a25d28532df204d5f61d2aaabc14a67b24096511e9cd393f31ea5787ca12: Status 404 returned error can't find the container with id 4814a25d28532df204d5f61d2aaabc14a67b24096511e9cd393f31ea5787ca12 Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.357018722Z" level=info msg="Ran pod sandbox 4814a25d28532df204d5f61d2aaabc14a67b24096511e9cd393f31ea5787ca12 with infra container: openshift-ingress/router-default-5cf6bbdff9-tnffs/POD" id=4fd824c4-281f-4b66-8101-8678d78afacb name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.364243249Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:bfd14188641a2d34e5b2b2ed3336b3b440155533f3fd7c6de776300dc3ff9e3b" id=87f40cda-9b8b-4418-bd18-79cf39a58e9c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.364484413Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:ad00ebc82f5cc671785b8401477b42db226d1c335ea60a4bb39bf28a961de840,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:bfd14188641a2d34e5b2b2ed3336b3b440155533f3fd7c6de776300dc3ff9e3b],Size_:411774965,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=87f40cda-9b8b-4418-bd18-79cf39a58e9c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:10.369360 2588 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-multus/network-metrics-daemon-hq594] Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.372953131Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:bfd14188641a2d34e5b2b2ed3336b3b440155533f3fd7c6de776300dc3ff9e3b" id=dfc76e09-ae63-4d4b-9ee1-bb1506097269 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.373166293Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:ad00ebc82f5cc671785b8401477b42db226d1c335ea60a4bb39bf28a961de840,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:bfd14188641a2d34e5b2b2ed3336b3b440155533f3fd7c6de776300dc3ff9e3b],Size_:411774965,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=dfc76e09-ae63-4d4b-9ee1-bb1506097269 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.374632693Z" level=info msg="Creating container: openshift-ingress/router-default-5cf6bbdff9-tnffs/router" id=cea6752c-d23e-44cf-b676-6421df1938b4 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.374756239Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:42:10Z [verbose] Add: openshift-multus:network-metrics-daemon-hq594:50152296-9489-4fa8-aa42-7858debe1c08:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/5a7d8972-1e35-45cd-add2-715bbc661172"}],"ips":[{"version":"4","interface":0,"address":"10.131.0.6/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: I0105 09:42:10.338984 5886 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-multus", Name:"network-metrics-daemon-hq594", UID:"50152296-9489-4fa8-aa42-7858debe1c08", APIVersion:"v1", ResourceVersion:"96783", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.6/23] from openshift-sdn Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.378273092Z" level=info msg="Got pod network &{Name:network-metrics-daemon-hq594 Namespace:openshift-multus ID:44ef9ac88935fc68d10995ab1204d3f92afbedeae9fabaf8c0abb737181fa9fd UID:50152296-9489-4fa8-aa42-7858debe1c08 NetNS:/var/run/netns/5a7d8972-1e35-45cd-add2-715bbc661172 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.378412426Z" level=info msg="Checking pod openshift-multus_network-metrics-daemon-hq594 for CNI network multus-cni-network (type=multus)" Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:42:10.404654 2588 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50152296_9489_4fa8_aa42_7858debe1c08.slice/crio-44ef9ac88935fc68d10995ab1204d3f92afbedeae9fabaf8c0abb737181fa9fd.scope WatchSource:0}: Error finding container 44ef9ac88935fc68d10995ab1204d3f92afbedeae9fabaf8c0abb737181fa9fd: Status 404 returned error can't find the container with id 44ef9ac88935fc68d10995ab1204d3f92afbedeae9fabaf8c0abb737181fa9fd Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.413515064Z" level=info msg="Ran pod sandbox 44ef9ac88935fc68d10995ab1204d3f92afbedeae9fabaf8c0abb737181fa9fd with infra container: openshift-multus/network-metrics-daemon-hq594/POD" id=bdfb661b-c8b0-4ca4-88ff-e93c5b94ae90 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.416072874Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9e33552d1aaea9cf6a301fa438cfc8305826c5ddb58ac8437d884e55e71c6e44" id=5fd0b588-61ad-40d6-81fa-618e5ccbdee1 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.416283708Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:7b87fc786118817b8913f28a61e01db37b08380f754bbaacf425508b208632a0,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9e33552d1aaea9cf6a301fa438cfc8305826c5ddb58ac8437d884e55e71c6e44],Size_:312894682,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=5fd0b588-61ad-40d6-81fa-618e5ccbdee1 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.419757666Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9e33552d1aaea9cf6a301fa438cfc8305826c5ddb58ac8437d884e55e71c6e44" id=57ec9a80-1a90-4df1-8327-484b61f30e0f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.419961173Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:7b87fc786118817b8913f28a61e01db37b08380f754bbaacf425508b208632a0,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9e33552d1aaea9cf6a301fa438cfc8305826c5ddb58ac8437d884e55e71c6e44],Size_:312894682,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=57ec9a80-1a90-4df1-8327-484b61f30e0f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.426959509Z" level=info msg="Creating container: openshift-multus/network-metrics-daemon-hq594/network-metrics-daemon" id=fba7301e-fd04-4190-b157-a1ab9b7474c5 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.427062077Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:10.433924 2588 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-ingress-canary/ingress-canary-8xrbh] Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:10.435887 2588 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/alertmanager-main-0] Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:42:10Z [verbose] Add: openshift-ingress-canary:ingress-canary-8xrbh:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/db10749b-0af7-4faa-bbc1-64769e589a18"}],"ips":[{"version":"4","interface":0,"address":"10.131.0.4/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: I0105 09:42:10.401209 5933 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-ingress-canary", Name:"ingress-canary-8xrbh", UID:"e04a47b4-aa8e-4d4e-9d45-aa8037dcb748", APIVersion:"v1", ResourceVersion:"96799", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.4/23] from openshift-sdn Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.435123397Z" level=info msg="Got pod network &{Name:ingress-canary-8xrbh Namespace:openshift-ingress-canary ID:2f42fdd8caa6c772e33c09307d855f2c6b126d16a962580bfad9af47fa893ef2 UID:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748 NetNS:/var/run/netns/db10749b-0af7-4faa-bbc1-64769e589a18 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.435328397Z" level=info msg="Checking pod openshift-ingress-canary_ingress-canary-8xrbh for CNI network multus-cni-network (type=multus)" Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:42:10Z [verbose] Add: openshift-monitoring:alertmanager-main-0:05c9a6ea-140a-49a2-bde5-f220b05aa252:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/c7a80f78-1128-4e9e-9931-f7ed057df8ad"}],"ips":[{"version":"4","interface":0,"address":"10.131.0.2/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: I0105 09:42:10.405819 5880 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"alertmanager-main-0", UID:"05c9a6ea-140a-49a2-bde5-f220b05aa252", APIVersion:"v1", ResourceVersion:"96794", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.2/23] from openshift-sdn Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.436771950Z" level=info msg="Got pod network &{Name:alertmanager-main-0 Namespace:openshift-monitoring ID:d6d0c82bc9774e51559f70121d26a57c7c38cce6fb83f23b63a8bb671c408015 UID:05c9a6ea-140a-49a2-bde5-f220b05aa252 NetNS:/var/run/netns/c7a80f78-1128-4e9e-9931-f7ed057df8ad Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.436882461Z" level=info msg="Checking pod openshift-monitoring_alertmanager-main-0 for CNI network multus-cni-network (type=multus)" Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.466354722Z" level=info msg="Ran pod sandbox 2f42fdd8caa6c772e33c09307d855f2c6b126d16a962580bfad9af47fa893ef2 with infra container: openshift-ingress-canary/ingress-canary-8xrbh/POD" id=f2cc8080-3b7a-4eec-9f06-f162f82a9cba name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.469206936Z" level=info msg="Ran pod sandbox d6d0c82bc9774e51559f70121d26a57c7c38cce6fb83f23b63a8bb671c408015 with infra container: openshift-monitoring/alertmanager-main-0/POD" id=a553a87e-65f2-44c7-9aec-483691caf6f0 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.481598099Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:5974989cda2655deb3052d601e7406eaf3caee0e7999a6eff251760aac19c199" id=e37fc4bf-f969-4327-a3b3-cbf4fa6ed0da name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.481768765Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:c34c97b229e46faa94bab6ad2559fd683bd036df06031c9097dd27f7daaf96c7,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:5974989cda2655deb3052d601e7406eaf3caee0e7999a6eff251760aac19c199],Size_:417780553,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=e37fc4bf-f969-4327-a3b3-cbf4fa6ed0da name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.481927630Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e6173b5d77def9e8058c7ff82f1ad65f537d1a81f88fcdabe6d8e30bff49ca89" id=92bd9fef-7e93-4d9d-bad9-782f1cfa0283 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.482042248Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:d168c738461c25cc0c3fa96cfcdf2dc337ca08f5bf3a6b550fcb29a59ea39270,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e6173b5d77def9e8058c7ff82f1ad65f537d1a81f88fcdabe6d8e30bff49ca89],Size_:333780768,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=92bd9fef-7e93-4d9d-bad9-782f1cfa0283 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.508912245Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:5974989cda2655deb3052d601e7406eaf3caee0e7999a6eff251760aac19c199" id=b033bffc-cc0d-460a-a372-c71e7a7c295b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.509118988Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:c34c97b229e46faa94bab6ad2559fd683bd036df06031c9097dd27f7daaf96c7,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:5974989cda2655deb3052d601e7406eaf3caee0e7999a6eff251760aac19c199],Size_:417780553,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b033bffc-cc0d-460a-a372-c71e7a7c295b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.517238593Z" level=info msg="Creating container: openshift-ingress-canary/ingress-canary-8xrbh/serve-healthcheck-canary" id=e6006e71-0a71-43b5-852e-de0153211e99 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.517365666Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.536945800Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e6173b5d77def9e8058c7ff82f1ad65f537d1a81f88fcdabe6d8e30bff49ca89" id=f7fdecb7-f197-4d44-90b8-ff0ce490418b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.557529990Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:d168c738461c25cc0c3fa96cfcdf2dc337ca08f5bf3a6b550fcb29a59ea39270,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e6173b5d77def9e8058c7ff82f1ad65f537d1a81f88fcdabe6d8e30bff49ca89],Size_:333780768,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=f7fdecb7-f197-4d44-90b8-ff0ce490418b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.564785881Z" level=info msg="Creating container: openshift-monitoring/alertmanager-main-0/alertmanager" id=5b732286-8214-4fd3-a11a-3ee8a45f6cf4 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.564894452Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.628381405Z" level=info msg="Created container 8eaf1cd2c43d71002a86869a186ab94e90ccc29be62720e01c85c11de8ae9285: openshift-ingress/router-default-5cf6bbdff9-tnffs/router" id=cea6752c-d23e-44cf-b676-6421df1938b4 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.630855412Z" level=info msg="Starting container: 8eaf1cd2c43d71002a86869a186ab94e90ccc29be62720e01c85c11de8ae9285" id=da88320e-0a85-452a-8643-3a20e9848275 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.653965495Z" level=info msg="Started container" PID=6166 containerID=8eaf1cd2c43d71002a86869a186ab94e90ccc29be62720e01c85c11de8ae9285 description=openshift-ingress/router-default-5cf6bbdff9-tnffs/router id=da88320e-0a85-452a-8643-3a20e9848275 name=/runtime.v1.RuntimeService/StartContainer sandboxID=4814a25d28532df204d5f61d2aaabc14a67b24096511e9cd393f31ea5787ca12 Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:10.731026 2588 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-machine-config-operator_machine-config-daemon-hf8f5_997b7ab4-939e-465c-9c7d-4a2ebe3a797a/machine-config-daemon/3.log" Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.772445118Z" level=info msg="Created container 8051e560b2b9ac3fbb8de3de8bf996d0cfd4b888017e0e97fda9c2c8d47e3f9a: openshift-multus/network-metrics-daemon-hq594/network-metrics-daemon" id=fba7301e-fd04-4190-b157-a1ab9b7474c5 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.773155390Z" level=info msg="Starting container: 8051e560b2b9ac3fbb8de3de8bf996d0cfd4b888017e0e97fda9c2c8d47e3f9a" id=39090632-f6ab-4b7e-8475-3668df8b2fad name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:10.801737 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-2pfzf" Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:10.803452 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:10.803565 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.804219630Z" level=info msg="Running pod sandbox: openshift-monitoring/telemeter-client-5598c645c4-mftdv/POD" id=2ec08c8a-4d69-4636-b273-4407db71b484 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.804285732Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.804461260Z" level=info msg="Running pod sandbox: openshift-network-diagnostics/network-check-target-xmq2g/POD" id=111a6daa-c3f9-41cd-8182-ea1f5080626b name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.804496269Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.804219485Z" level=info msg="Running pod sandbox: openshift-dns/dns-default-2pfzf/POD" id=0c59acbf-ab11-4bf8-ad7d-1aeb6b6c39d8 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.804824302Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.830157717Z" level=info msg="Created container 29d242ba69edfb0ca301e890f9169a7355a0889dbda9e344277448360467524b: openshift-monitoring/alertmanager-main-0/alertmanager" id=5b732286-8214-4fd3-a11a-3ee8a45f6cf4 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.834137644Z" level=info msg="Starting container: 29d242ba69edfb0ca301e890f9169a7355a0889dbda9e344277448360467524b" id=fcf26911-6bef-4544-88ee-cfb1ae95b6d9 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.851817195Z" level=info msg="Started container" PID=6215 containerID=8051e560b2b9ac3fbb8de3de8bf996d0cfd4b888017e0e97fda9c2c8d47e3f9a description=openshift-multus/network-metrics-daemon-hq594/network-metrics-daemon id=39090632-f6ab-4b7e-8475-3668df8b2fad name=/runtime.v1.RuntimeService/StartContainer sandboxID=44ef9ac88935fc68d10995ab1204d3f92afbedeae9fabaf8c0abb737181fa9fd Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.874132581Z" level=info msg="Started container" PID=6234 containerID=29d242ba69edfb0ca301e890f9169a7355a0889dbda9e344277448360467524b description=openshift-monitoring/alertmanager-main-0/alertmanager id=fcf26911-6bef-4544-88ee-cfb1ae95b6d9 name=/runtime.v1.RuntimeService/StartContainer sandboxID=d6d0c82bc9774e51559f70121d26a57c7c38cce6fb83f23b63a8bb671c408015 Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.888747114Z" level=info msg="Created container 90eebd6c0e12b30f0ef5186e6b2d5c9daec4715fe29ad932724cad4a69ee55b4: openshift-ingress-canary/ingress-canary-8xrbh/serve-healthcheck-canary" id=e6006e71-0a71-43b5-852e-de0153211e99 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.889515178Z" level=info msg="Starting container: 90eebd6c0e12b30f0ef5186e6b2d5c9daec4715fe29ad932724cad4a69ee55b4" id=c06c2cdf-aa62-42a6-b003-3ccccc3a83c2 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.908068896Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=813a01e0-d45b-46dd-9622-0c16edda767d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.912387379Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=17091bad-147a-4907-9ef5-eacbaa7456c5 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.917627552Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=813a01e0-d45b-46dd-9622-0c16edda767d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.917828618Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=17091bad-147a-4907-9ef5-eacbaa7456c5 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.919457899Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=c28c2219-fec9-4c7a-9b0d-cc73e3815bb8 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.919541881Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=ed51e3a6-c684-449c-8d40-02241ecb6900 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.919643665Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=c28c2219-fec9-4c7a-9b0d-cc73e3815bb8 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.919703408Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=ed51e3a6-c684-449c-8d40-02241ecb6900 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.920547173Z" level=info msg="Creating container: openshift-multus/network-metrics-daemon-hq594/kube-rbac-proxy" id=753329d3-53ea-4c39-96b9-efc490ecc353 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.920662244Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.926918352Z" level=info msg="Creating container: openshift-monitoring/alertmanager-main-0/config-reloader" id=154bc56b-8433-4596-a49a-3d0a39b33d4c name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.927017740Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:10.928466461Z" level=info msg="Started container" PID=6221 containerID=90eebd6c0e12b30f0ef5186e6b2d5c9daec4715fe29ad932724cad4a69ee55b4 description=openshift-ingress-canary/ingress-canary-8xrbh/serve-healthcheck-canary id=c06c2cdf-aa62-42a6-b003-3ccccc3a83c2 name=/runtime.v1.RuntimeService/StartContainer sandboxID=2f42fdd8caa6c772e33c09307d855f2c6b126d16a962580bfad9af47fa893ef2 Jan 05 09:42:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:10.975103 2588 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-machine-config-operator_machine-config-daemon-hf8f5_997b7ab4-939e-465c-9c7d-4a2ebe3a797a/oauth-proxy/3.log" Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.052770262Z" level=info msg="Got pod network &{Name:dns-default-2pfzf Namespace:openshift-dns ID:4f0a921bd0d3c8f309751665c310e818034ecd3252406f2a101ca811d04377c5 UID:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 NetNS:/var/run/netns/7b20456c-f122-4dc5-ba26-66ea4aeb44a5 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.052808727Z" level=info msg="Adding pod openshift-dns_dns-default-2pfzf to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.067910839Z" level=info msg="Got pod network &{Name:network-check-target-xmq2g Namespace:openshift-network-diagnostics ID:df03e4f874ec90019c812fc416c2d12340d64e387a2a542bdc3beee879bc31bd UID:f364a949-09d7-4cfb-83ff-e532b822a557 NetNS:/var/run/netns/13c36503-2dc5-4ab8-a8bd-dd131abd1770 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.067938170Z" level=info msg="Adding pod openshift-network-diagnostics_network-check-target-xmq2g to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.074672831Z" level=info msg="Got pod network &{Name:telemeter-client-5598c645c4-mftdv Namespace:openshift-monitoring ID:f9cc29d007bd3c948c46f781abd08afdb46359ec50fc31f8c503fcdb35b7aa22 UID:fc45e847-8f55-4c48-8c48-4bfee2419dca NetNS:/var/run/netns/cac8a001-4260-4cae-ad1e-0de0816fe057 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.074710051Z" level=info msg="Adding pod openshift-monitoring_telemeter-client-5598c645c4-mftdv to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:11.295657 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-8xrbh" event=&{ID:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748 Type:ContainerStarted Data:90eebd6c0e12b30f0ef5186e6b2d5c9daec4715fe29ad932724cad4a69ee55b4} Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:11.295699 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-8xrbh" event=&{ID:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748 Type:ContainerStarted Data:2f42fdd8caa6c772e33c09307d855f2c6b126d16a962580bfad9af47fa893ef2} Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:11.307928 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event=&{ID:05c9a6ea-140a-49a2-bde5-f220b05aa252 Type:ContainerStarted Data:29d242ba69edfb0ca301e890f9169a7355a0889dbda9e344277448360467524b} Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:11.307963 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event=&{ID:05c9a6ea-140a-49a2-bde5-f220b05aa252 Type:ContainerStarted Data:d6d0c82bc9774e51559f70121d26a57c7c38cce6fb83f23b63a8bb671c408015} Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:11.309074 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" event=&{ID:c3b10ff4-a243-4a4e-b27d-f613e258666d Type:ContainerStarted Data:8eaf1cd2c43d71002a86869a186ab94e90ccc29be62720e01c85c11de8ae9285} Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:11.309107 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" event=&{ID:c3b10ff4-a243-4a4e-b27d-f613e258666d Type:ContainerStarted Data:4814a25d28532df204d5f61d2aaabc14a67b24096511e9cd393f31ea5787ca12} Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:11.312797 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hq594" event=&{ID:50152296-9489-4fa8-aa42-7858debe1c08 Type:ContainerStarted Data:8051e560b2b9ac3fbb8de3de8bf996d0cfd4b888017e0e97fda9c2c8d47e3f9a} Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:11.312824 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hq594" event=&{ID:50152296-9489-4fa8-aa42-7858debe1c08 Type:ContainerStarted Data:44ef9ac88935fc68d10995ab1204d3f92afbedeae9fabaf8c0abb737181fa9fd} Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:42:11Z [verbose] Add: openshift-monitoring:prometheus-operator-admission-webhook-598468bb49-94dsn:88745ab7-efe7-42c0-a2e8-75591518333d:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/67866f67-3eb9-4569-aa7e-69d12f039663"}],"ips":[{"version":"4","interface":0,"address":"10.131.0.5/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: I0105 09:42:11.309676 5905 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"prometheus-operator-admission-webhook-598468bb49-94dsn", UID:"88745ab7-efe7-42c0-a2e8-75591518333d", APIVersion:"v1", ResourceVersion:"96801", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.5/23] from openshift-sdn Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.416040278Z" level=info msg="Got pod network &{Name:prometheus-operator-admission-webhook-598468bb49-94dsn Namespace:openshift-monitoring ID:e10e02d14faf7624d78743e0d1899ddf64963db130ecf653b0849bcde995a72c UID:88745ab7-efe7-42c0-a2e8-75591518333d NetNS:/var/run/netns/67866f67-3eb9-4569-aa7e-69d12f039663 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.416225796Z" level=info msg="Checking pod openshift-monitoring_prometheus-operator-admission-webhook-598468bb49-94dsn for CNI network multus-cni-network (type=multus)" Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:11.439967 2588 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn] Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:11.451112 2588 kubelet.go:2229] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:42:11.451293 2588 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88745ab7_efe7_42c0_a2e8_75591518333d.slice/crio-e10e02d14faf7624d78743e0d1899ddf64963db130ecf653b0849bcde995a72c.scope WatchSource:0}: Error finding container e10e02d14faf7624d78743e0d1899ddf64963db130ecf653b0849bcde995a72c: Status 404 returned error can't find the container with id e10e02d14faf7624d78743e0d1899ddf64963db130ecf653b0849bcde995a72c Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:11.460640 2588 patch_prober.go:29] interesting pod/router-default-5cf6bbdff9-tnffs container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: [-]has-synced failed: reason withheld Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: [+]process-running ok Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: healthz check failed Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:11.460702 2588 prober.go:114] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" podUID=c3b10ff4-a243-4a4e-b27d-f613e258666d containerName="router" probeResult=failure output="HTTP probe failed with statuscode: 500" Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.462401040Z" level=info msg="Ran pod sandbox e10e02d14faf7624d78743e0d1899ddf64963db130ecf653b0849bcde995a72c with infra container: openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn/POD" id=08d00f52-7155-4997-9c76-ed8b4000bc50 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.492514048Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:dbf46164dcc3b2b6ef5df5d043ddc4fd859fd2c73c4862808424b3df6a1e27e7" id=4f90cbc6-438c-4852-ab9a-151db10a5c84 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.492809016Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:20e9403b34a0d9d25151a4ddbee3f69b99894fd8e058e6cb7310fe95080ecb03,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:dbf46164dcc3b2b6ef5df5d043ddc4fd859fd2c73c4862808424b3df6a1e27e7],Size_:374119253,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=4f90cbc6-438c-4852-ab9a-151db10a5c84 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.501671318Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:dbf46164dcc3b2b6ef5df5d043ddc4fd859fd2c73c4862808424b3df6a1e27e7" id=7f1cf750-b4a3-4cc6-97be-3143c38da62b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.501939560Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:20e9403b34a0d9d25151a4ddbee3f69b99894fd8e058e6cb7310fe95080ecb03,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:dbf46164dcc3b2b6ef5df5d043ddc4fd859fd2c73c4862808424b3df6a1e27e7],Size_:374119253,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=7f1cf750-b4a3-4cc6-97be-3143c38da62b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.507371476Z" level=info msg="Creating container: openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn/prometheus-operator-admission-webhook" id=98934f80-6f6f-4622-ac58-ecf805ec0b39 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.507505304Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.568131798Z" level=info msg="Created container df204cb60308dd066ca32f1f38b7a71d29064341d03539fedabfb840583c8a96: openshift-monitoring/alertmanager-main-0/config-reloader" id=154bc56b-8433-4596-a49a-3d0a39b33d4c name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.568681344Z" level=info msg="Starting container: df204cb60308dd066ca32f1f38b7a71d29064341d03539fedabfb840583c8a96" id=1078a6fe-3036-42ff-9982-3e57d09d9f1e name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.610884557Z" level=info msg="Started container" PID=6397 containerID=df204cb60308dd066ca32f1f38b7a71d29064341d03539fedabfb840583c8a96 description=openshift-monitoring/alertmanager-main-0/config-reloader id=1078a6fe-3036-42ff-9982-3e57d09d9f1e name=/runtime.v1.RuntimeService/StartContainer sandboxID=d6d0c82bc9774e51559f70121d26a57c7c38cce6fb83f23b63a8bb671c408015 Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.623639459Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=5be108b2-0171-4a9b-a652-3f309c2bb387 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.623884920Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=5be108b2-0171-4a9b-a652-3f309c2bb387 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.634551764Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=f6f85525-32fc-4e22-b939-f37e4657b40e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.634763888Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=f6f85525-32fc-4e22-b939-f37e4657b40e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.635849322Z" level=info msg="Creating container: openshift-monitoring/alertmanager-main-0/alertmanager-proxy" id=b0c2e644-4505-4301-8237-95b8d221af59 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.635984481Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:11.794732 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:11.795387 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.798198881Z" level=info msg="Running pod sandbox: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/POD" id=e91c6a60-ff52-4909-a60c-3d6c50f02a0c name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.798255897Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.800720842Z" level=info msg="Running pod sandbox: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/POD" id=fbf40af3-df42-4b07-a1a3-974b71a0b0e1 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.800761891Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.802902691Z" level=info msg="Created container 51706a2d3e92644b7041be24aad6900f3dea2028777516e28cabf328c9253df9: openshift-multus/network-metrics-daemon-hq594/kube-rbac-proxy" id=753329d3-53ea-4c39-96b9-efc490ecc353 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.803279502Z" level=info msg="Starting container: 51706a2d3e92644b7041be24aad6900f3dea2028777516e28cabf328c9253df9" id=6fe4e2e5-18d0-40bf-ad26-e5149141fb6b name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.889236710Z" level=info msg="Started container" PID=6475 containerID=51706a2d3e92644b7041be24aad6900f3dea2028777516e28cabf328c9253df9 description=openshift-multus/network-metrics-daemon-hq594/kube-rbac-proxy id=6fe4e2e5-18d0-40bf-ad26-e5149141fb6b name=/runtime.v1.RuntimeService/StartContainer sandboxID=44ef9ac88935fc68d10995ab1204d3f92afbedeae9fabaf8c0abb737181fa9fd Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.890916873Z" level=info msg="Got pod network &{Name:kube-state-metrics-6798b548b9-xl88q Namespace:openshift-monitoring ID:1e97b2e64ee6d4f76d829e9619da041f24e571e3de5454c48314d411b8770c02 UID:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e NetNS:/var/run/netns/cbe37ad3-4a92-423a-8f3c-dda00e586a03 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.890950034Z" level=info msg="Adding pod openshift-monitoring_kube-state-metrics-6798b548b9-xl88q to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.892036835Z" level=info msg="Got pod network &{Name:openshift-state-metrics-6f5dcdf79b-wq256 Namespace:openshift-monitoring ID:7a37710cfebf59892a4d6b3c2e12765b60414f6a7cfb8933a050df609c94e690 UID:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 NetNS:/var/run/netns/e7429438-8b02-46d6-a53c-f91c31850d37 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:42:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:11.892065645Z" level=info msg="Adding pod openshift-monitoring_openshift-state-metrics-6f5dcdf79b-wq256 to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.081341425Z" level=info msg="Created container 85c843961a0e8a2283e240b3c350b3d3823e13fbe3cb8f291a0e809df84275e7: openshift-monitoring/alertmanager-main-0/alertmanager-proxy" id=b0c2e644-4505-4301-8237-95b8d221af59 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.093957183Z" level=info msg="Created container 0b999d4ad314ba518c86d2e8df430b23bade34adf813f2325e76c8d1d79e1b52: openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn/prometheus-operator-admission-webhook" id=98934f80-6f6f-4622-ac58-ecf805ec0b39 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.102863650Z" level=info msg="Starting container: 85c843961a0e8a2283e240b3c350b3d3823e13fbe3cb8f291a0e809df84275e7" id=c0d8eeac-d5d1-414c-a1d4-4be8291d96e5 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.102870461Z" level=info msg="Starting container: 0b999d4ad314ba518c86d2e8df430b23bade34adf813f2325e76c8d1d79e1b52" id=f1956d68-0bf1-438e-b9c3-f14ada1a67fb name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.171075273Z" level=info msg="Started container" PID=6501 containerID=85c843961a0e8a2283e240b3c350b3d3823e13fbe3cb8f291a0e809df84275e7 description=openshift-monitoring/alertmanager-main-0/alertmanager-proxy id=c0d8eeac-d5d1-414c-a1d4-4be8291d96e5 name=/runtime.v1.RuntimeService/StartContainer sandboxID=d6d0c82bc9774e51559f70121d26a57c7c38cce6fb83f23b63a8bb671c408015 Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.198847564Z" level=info msg="Started container" PID=6502 containerID=0b999d4ad314ba518c86d2e8df430b23bade34adf813f2325e76c8d1d79e1b52 description=openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn/prometheus-operator-admission-webhook id=f1956d68-0bf1-438e-b9c3-f14ada1a67fb name=/runtime.v1.RuntimeService/StartContainer sandboxID=e10e02d14faf7624d78743e0d1899ddf64963db130ecf653b0849bcde995a72c Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.222486311Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=d536c102-a173-4da2-b4f5-2dee038f0324 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.222722338Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=d536c102-a173-4da2-b4f5-2dee038f0324 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.225831832Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=4b476d5e-26ef-4c31-b6fd-0c1c80fec805 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.226053524Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=4b476d5e-26ef-4c31-b6fd-0c1c80fec805 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.227088217Z" level=info msg="Creating container: openshift-monitoring/alertmanager-main-0/kube-rbac-proxy" id=933381b6-f5a7-43c9-aaee-fd8783dfb50f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.227219819Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:12.232150 2588 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l] Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:42:12Z [verbose] Add: openshift-monitoring:thanos-querier-6f4f5fb7cf-rdm9l:7c63c9be-d1ad-44b6-8b33-09af3ee314af:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/d054aa63-f5ac-4893-88ff-2a7d67c9ea25"}],"ips":[{"version":"4","interface":0,"address":"10.131.0.7/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: I0105 09:42:12.104129 6063 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"thanos-querier-6f4f5fb7cf-rdm9l", UID:"7c63c9be-d1ad-44b6-8b33-09af3ee314af", APIVersion:"v1", ResourceVersion:"96806", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.7/23] from openshift-sdn Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.249698941Z" level=info msg="Got pod network &{Name:thanos-querier-6f4f5fb7cf-rdm9l Namespace:openshift-monitoring ID:fcdbc115e8986f7ba4e1fc9bfa0728fd149a52733e5f0956b91d0b8262b33ca6 UID:7c63c9be-d1ad-44b6-8b33-09af3ee314af NetNS:/var/run/netns/d054aa63-f5ac-4893-88ff-2a7d67c9ea25 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.249851783Z" level=info msg="Checking pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l for CNI network multus-cni-network (type=multus)" Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:42:12.297554 2588 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c63c9be_d1ad_44b6_8b33_09af3ee314af.slice/crio-fcdbc115e8986f7ba4e1fc9bfa0728fd149a52733e5f0956b91d0b8262b33ca6.scope WatchSource:0}: Error finding container fcdbc115e8986f7ba4e1fc9bfa0728fd149a52733e5f0956b91d0b8262b33ca6: Status 404 returned error can't find the container with id fcdbc115e8986f7ba4e1fc9bfa0728fd149a52733e5f0956b91d0b8262b33ca6 Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.298394916Z" level=info msg="Ran pod sandbox fcdbc115e8986f7ba4e1fc9bfa0728fd149a52733e5f0956b91d0b8262b33ca6 with infra container: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/POD" id=d5df9f78-035a-4776-a598-68baf0ff7872 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.303675794Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc" id=c27c31d6-2a3f-45da-a7a8-98c35c1edc21 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.303897097Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:08ff7a43d9185529520a29e114f0c28b070000b464a44b347240015f3b466488,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc],Size_:368278089,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=c27c31d6-2a3f-45da-a7a8-98c35c1edc21 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.305176348Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc" id=cf5c4a07-b87a-4e79-9aa9-80c0b5d033ff name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.305435939Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:08ff7a43d9185529520a29e114f0c28b070000b464a44b347240015f3b466488,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc],Size_:368278089,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=cf5c4a07-b87a-4e79-9aa9-80c0b5d033ff name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.306584961Z" level=info msg="Creating container: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/thanos-query" id=c138be33-de79-42f7-8664-13ebed2b20a9 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.306700407Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:12.316820 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" event=&{ID:7c63c9be-d1ad-44b6-8b33-09af3ee314af Type:ContainerStarted Data:fcdbc115e8986f7ba4e1fc9bfa0728fd149a52733e5f0956b91d0b8262b33ca6} Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:12.318616 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event=&{ID:05c9a6ea-140a-49a2-bde5-f220b05aa252 Type:ContainerStarted Data:85c843961a0e8a2283e240b3c350b3d3823e13fbe3cb8f291a0e809df84275e7} Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:12.318645 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event=&{ID:05c9a6ea-140a-49a2-bde5-f220b05aa252 Type:ContainerStarted Data:df204cb60308dd066ca32f1f38b7a71d29064341d03539fedabfb840583c8a96} Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:12.320148 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hq594" event=&{ID:50152296-9489-4fa8-aa42-7858debe1c08 Type:ContainerStarted Data:51706a2d3e92644b7041be24aad6900f3dea2028777516e28cabf328c9253df9} Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:12.325067 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" event=&{ID:88745ab7-efe7-42c0-a2e8-75591518333d Type:ContainerStarted Data:0b999d4ad314ba518c86d2e8df430b23bade34adf813f2325e76c8d1d79e1b52} Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:12.325103 2588 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:12.325124 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" event=&{ID:88745ab7-efe7-42c0-a2e8-75591518333d Type:ContainerStarted Data:e10e02d14faf7624d78743e0d1899ddf64963db130ecf653b0849bcde995a72c} Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:12.326992 2588 patch_prober.go:29] interesting pod/prometheus-operator-admission-webhook-598468bb49-94dsn container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.131.0.5:8443/healthz\": dial tcp 10.131.0.5:8443: connect: connection refused" start-of-body= Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:12.327048 2588 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" podUID=88745ab7-efe7-42c0-a2e8-75591518333d containerName="prometheus-operator-admission-webhook" probeResult=failure output="Get \"https://10.131.0.5:8443/healthz\": dial tcp 10.131.0.5:8443: connect: connection refused" Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:12.444109 2588 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:12.446994 2588 patch_prober.go:29] interesting pod/router-default-5cf6bbdff9-tnffs container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: [-]has-synced failed: reason withheld Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: [+]process-running ok Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: healthz check failed Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:12.447042 2588 prober.go:114] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" podUID=c3b10ff4-a243-4a4e-b27d-f613e258666d containerName="router" probeResult=failure output="HTTP probe failed with statuscode: 500" Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.494620673Z" level=info msg="Created container 53a26d2c619d3ef8230bda385bc9245dc43ada80f94fe54f1f79b82763d2f7a8: openshift-monitoring/alertmanager-main-0/kube-rbac-proxy" id=933381b6-f5a7-43c9-aaee-fd8783dfb50f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.495206942Z" level=info msg="Starting container: 53a26d2c619d3ef8230bda385bc9245dc43ada80f94fe54f1f79b82763d2f7a8" id=a3b533a3-f3a0-408a-a58b-77c80e00978a name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.512248080Z" level=info msg="Started container" PID=6684 containerID=53a26d2c619d3ef8230bda385bc9245dc43ada80f94fe54f1f79b82763d2f7a8 description=openshift-monitoring/alertmanager-main-0/kube-rbac-proxy id=a3b533a3-f3a0-408a-a58b-77c80e00978a name=/runtime.v1.RuntimeService/StartContainer sandboxID=d6d0c82bc9774e51559f70121d26a57c7c38cce6fb83f23b63a8bb671c408015 Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.530357922Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=d62db9b3-46a1-4828-a51e-35300bf23ce4 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.530585960Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=d62db9b3-46a1-4828-a51e-35300bf23ce4 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.531817822Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=81329b61-0fa4-4482-9bc9-134f0f69194c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.532014178Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=81329b61-0fa4-4482-9bc9-134f0f69194c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.533473698Z" level=info msg="Creating container: openshift-monitoring/alertmanager-main-0/kube-rbac-proxy-metric" id=9286128b-e2f5-4fdd-a95b-6987484b4d89 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.533600991Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:12.567336 2588 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/telemeter-client-5598c645c4-mftdv] Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:42:12Z [verbose] Add: openshift-monitoring:telemeter-client-5598c645c4-mftdv:fc45e847-8f55-4c48-8c48-4bfee2419dca:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/cac8a001-4260-4cae-ad1e-0de0816fe057"}],"ips":[{"version":"4","interface":0,"address":"10.131.0.10/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: I0105 09:42:12.436723 6324 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"telemeter-client-5598c645c4-mftdv", UID:"fc45e847-8f55-4c48-8c48-4bfee2419dca", APIVersion:"v1", ResourceVersion:"96789", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.10/23] from openshift-sdn Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.584290644Z" level=info msg="Got pod network &{Name:telemeter-client-5598c645c4-mftdv Namespace:openshift-monitoring ID:f9cc29d007bd3c948c46f781abd08afdb46359ec50fc31f8c503fcdb35b7aa22 UID:fc45e847-8f55-4c48-8c48-4bfee2419dca NetNS:/var/run/netns/cac8a001-4260-4cae-ad1e-0de0816fe057 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.584441847Z" level=info msg="Checking pod openshift-monitoring_telemeter-client-5598c645c4-mftdv for CNI network multus-cni-network (type=multus)" Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:42:12.608946 2588 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc45e847_8f55_4c48_8c48_4bfee2419dca.slice/crio-f9cc29d007bd3c948c46f781abd08afdb46359ec50fc31f8c503fcdb35b7aa22.scope WatchSource:0}: Error finding container f9cc29d007bd3c948c46f781abd08afdb46359ec50fc31f8c503fcdb35b7aa22: Status 404 returned error can't find the container with id f9cc29d007bd3c948c46f781abd08afdb46359ec50fc31f8c503fcdb35b7aa22 Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.611661551Z" level=info msg="Ran pod sandbox f9cc29d007bd3c948c46f781abd08afdb46359ec50fc31f8c503fcdb35b7aa22 with infra container: openshift-monitoring/telemeter-client-5598c645c4-mftdv/POD" id=2ec08c8a-4d69-4636-b273-4407db71b484 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.619341654Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:85f671344a3bc1584db4caed39939b6dd55644a5bd806e118527f31dac91252e" id=aa1a6678-c538-4213-9c6f-6eb4deb98b8d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.619596547Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b66b2c4519cdd77137d5eaa713199fe202373c0d1e029b31472b9fc8a68a2aa7,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:85f671344a3bc1584db4caed39939b6dd55644a5bd806e118527f31dac91252e],Size_:338794817,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=aa1a6678-c538-4213-9c6f-6eb4deb98b8d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.641927009Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:85f671344a3bc1584db4caed39939b6dd55644a5bd806e118527f31dac91252e" id=4896af1b-4e50-4e96-912b-26ec8e720c2e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.642154664Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b66b2c4519cdd77137d5eaa713199fe202373c0d1e029b31472b9fc8a68a2aa7,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:85f671344a3bc1584db4caed39939b6dd55644a5bd806e118527f31dac91252e],Size_:338794817,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=4896af1b-4e50-4e96-912b-26ec8e720c2e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.646107499Z" level=info msg="Creating container: openshift-monitoring/telemeter-client-5598c645c4-mftdv/telemeter-client" id=846edc2b-3e5a-4123-935e-229d91bf4322 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.646286854Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.649756726Z" level=info msg="Created container 578cce0ed1aa4795c3b6d8648ecf4b937233d871e6a027eb233bec20d4333c1f: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/thanos-query" id=c138be33-de79-42f7-8664-13ebed2b20a9 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.651836821Z" level=info msg="Starting container: 578cce0ed1aa4795c3b6d8648ecf4b937233d871e6a027eb233bec20d4333c1f" id=1a719fdf-cf73-40eb-8436-27ac28420964 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.696027333Z" level=info msg="Started container" PID=6712 containerID=578cce0ed1aa4795c3b6d8648ecf4b937233d871e6a027eb233bec20d4333c1f description=openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/thanos-query id=1a719fdf-cf73-40eb-8436-27ac28420964 name=/runtime.v1.RuntimeService/StartContainer sandboxID=fcdbc115e8986f7ba4e1fc9bfa0728fd149a52733e5f0956b91d0b8262b33ca6 Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.748954603Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=0977c201-be15-4792-b075-d0c2c584bec4 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.749218582Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=0977c201-be15-4792-b075-d0c2c584bec4 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.750454697Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=77eef02f-3375-46c6-8643-0ef678da1068 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.750634146Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=77eef02f-3375-46c6-8643-0ef678da1068 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.751964337Z" level=info msg="Creating container: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/oauth-proxy" id=8becde2f-45fb-41c9-9f31-695820ba74d3 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.752083490Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:12.783327 2588 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-dns/dns-default-2pfzf] Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:12.800244 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.800769036Z" level=info msg="Running pod sandbox: openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh/POD" id=a63a286d-5166-431d-a81d-ecfec74c8af7 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.800824608Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:42:12Z [verbose] Add: openshift-dns:dns-default-2pfzf:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/7b20456c-f122-4dc5-ba26-66ea4aeb44a5"}],"ips":[{"version":"4","interface":0,"address":"10.131.0.8/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: I0105 09:42:12.691334 6313 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-dns", Name:"dns-default-2pfzf", UID:"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63", APIVersion:"v1", ResourceVersion:"96785", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.8/23] from openshift-sdn Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.803595379Z" level=info msg="Got pod network &{Name:dns-default-2pfzf Namespace:openshift-dns ID:4f0a921bd0d3c8f309751665c310e818034ecd3252406f2a101ca811d04377c5 UID:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 NetNS:/var/run/netns/7b20456c-f122-4dc5-ba26-66ea4aeb44a5 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.803736771Z" level=info msg="Checking pod openshift-dns_dns-default-2pfzf for CNI network multus-cni-network (type=multus)" Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:42:12.840772 2588 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1cc2aadb_14f2_4d6e_97fc_6fdab8889a63.slice/crio-4f0a921bd0d3c8f309751665c310e818034ecd3252406f2a101ca811d04377c5.scope WatchSource:0}: Error finding container 4f0a921bd0d3c8f309751665c310e818034ecd3252406f2a101ca811d04377c5: Status 404 returned error can't find the container with id 4f0a921bd0d3c8f309751665c310e818034ecd3252406f2a101ca811d04377c5 Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.859534233Z" level=info msg="Ran pod sandbox 4f0a921bd0d3c8f309751665c310e818034ecd3252406f2a101ca811d04377c5 with infra container: openshift-dns/dns-default-2pfzf/POD" id=0c59acbf-ab11-4bf8-ad7d-1aeb6b6c39d8 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.864139362Z" level=info msg="Created container 4c970eddded2809e2e506cfb0b207a19b375bc49b3a1f9bfd58b202b7c1f6ebd: openshift-monitoring/alertmanager-main-0/kube-rbac-proxy-metric" id=9286128b-e2f5-4fdd-a95b-6987484b4d89 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.866750816Z" level=info msg="Starting container: 4c970eddded2809e2e506cfb0b207a19b375bc49b3a1f9bfd58b202b7c1f6ebd" id=152a3ef5-3c4b-4072-85aa-50abcc3b273d name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.874176377Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f674b597694b13b7060c9ee611a403f1f87a98e05682c33f9a86140061ee9d9e" id=5aaf7d28-e83b-4ed1-ae12-008e0e2c6574 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.894335688Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:cdb247afcbd9968f1a59680dcc4de0364f7ddd8fc25dd3a12db59f05c8e026a0,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f674b597694b13b7060c9ee611a403f1f87a98e05682c33f9a86140061ee9d9e],Size_:404212434,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=5aaf7d28-e83b-4ed1-ae12-008e0e2c6574 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.913661182Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f674b597694b13b7060c9ee611a403f1f87a98e05682c33f9a86140061ee9d9e" id=00938504-d07c-4017-8d02-c0a895125024 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.913881813Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:cdb247afcbd9968f1a59680dcc4de0364f7ddd8fc25dd3a12db59f05c8e026a0,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f674b597694b13b7060c9ee611a403f1f87a98e05682c33f9a86140061ee9d9e],Size_:404212434,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=00938504-d07c-4017-8d02-c0a895125024 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.932238527Z" level=info msg="Creating container: openshift-dns/dns-default-2pfzf/dns" id=8386fadb-94fb-4fa1-919b-71a124e20e93 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.932395170Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.935927579Z" level=info msg="Started container" PID=6787 containerID=4c970eddded2809e2e506cfb0b207a19b375bc49b3a1f9bfd58b202b7c1f6ebd description=openshift-monitoring/alertmanager-main-0/kube-rbac-proxy-metric id=152a3ef5-3c4b-4072-85aa-50abcc3b273d name=/runtime.v1.RuntimeService/StartContainer sandboxID=d6d0c82bc9774e51559f70121d26a57c7c38cce6fb83f23b63a8bb671c408015 Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.945868484Z" level=info msg="Got pod network &{Name:prometheus-adapter-5894cc86c7-qhmbh Namespace:openshift-monitoring ID:9d9807a6998df2507a4aa5b9bad79e347bc60d62b254989738c927929d827e6e UID:a40bf20d-67a4-45e7-b701-ca5a75376c1f NetNS:/var/run/netns/2bb8265e-9d3b-4130-ab12-f8f63e286c5d Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:42:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:12.945900865Z" level=info msg="Adding pod openshift-monitoring_prometheus-adapter-5894cc86c7-qhmbh to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.052864790Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816" id=bfe60990-ddf2-49d5-bdc2-bd12385935bf name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.053089972Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:955ee2b41ecd88e3a9c0183a0b65d6edbf524cefa1318837dcbc82b0466c4d55,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816],Size_:309576938,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=bfe60990-ddf2-49d5-bdc2-bd12385935bf name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.072253832Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816" id=e1b362db-d617-4c32-a145-1a397dfd3d46 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.072523607Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:955ee2b41ecd88e3a9c0183a0b65d6edbf524cefa1318837dcbc82b0466c4d55,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816],Size_:309576938,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=e1b362db-d617-4c32-a145-1a397dfd3d46 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.073600821Z" level=info msg="Creating container: openshift-monitoring/alertmanager-main-0/prom-label-proxy" id=fb4c36ac-0c9d-47ce-a14d-6c5c16f7aa3e name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.073719704Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:13.125458 2588 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-network-diagnostics/network-check-target-xmq2g] Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:42:13Z [verbose] Add: openshift-network-diagnostics:network-check-target-xmq2g:f364a949-09d7-4cfb-83ff-e532b822a557:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/13c36503-2dc5-4ab8-a8bd-dd131abd1770"}],"ips":[{"version":"4","interface":0,"address":"10.131.0.9/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: I0105 09:42:13.089428 6319 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-network-diagnostics", Name:"network-check-target-xmq2g", UID:"f364a949-09d7-4cfb-83ff-e532b822a557", APIVersion:"v1", ResourceVersion:"96749", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.9/23] from openshift-sdn Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.186363535Z" level=info msg="Got pod network &{Name:network-check-target-xmq2g Namespace:openshift-network-diagnostics ID:df03e4f874ec90019c812fc416c2d12340d64e387a2a542bdc3beee879bc31bd UID:f364a949-09d7-4cfb-83ff-e532b822a557 NetNS:/var/run/netns/13c36503-2dc5-4ab8-a8bd-dd131abd1770 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.186517302Z" level=info msg="Checking pod openshift-network-diagnostics_network-check-target-xmq2g for CNI network multus-cni-network (type=multus)" Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.188775663Z" level=info msg="Created container 2c1936a19655345a892373f86a623f7409e46829cbb1adeca75e2879d268aa08: openshift-monitoring/telemeter-client-5598c645c4-mftdv/telemeter-client" id=846edc2b-3e5a-4123-935e-229d91bf4322 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.198323828Z" level=info msg="Starting container: 2c1936a19655345a892373f86a623f7409e46829cbb1adeca75e2879d268aa08" id=5fa8b7ff-e67e-4f11-88a3-58e85f85f349 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:42:13.231586 2588 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf364a949_09d7_4cfb_83ff_e532b822a557.slice/crio-df03e4f874ec90019c812fc416c2d12340d64e387a2a542bdc3beee879bc31bd.scope WatchSource:0}: Error finding container df03e4f874ec90019c812fc416c2d12340d64e387a2a542bdc3beee879bc31bd: Status 404 returned error can't find the container with id df03e4f874ec90019c812fc416c2d12340d64e387a2a542bdc3beee879bc31bd Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.241722364Z" level=info msg="Ran pod sandbox df03e4f874ec90019c812fc416c2d12340d64e387a2a542bdc3beee879bc31bd with infra container: openshift-network-diagnostics/network-check-target-xmq2g/POD" id=111a6daa-c3f9-41cd-8182-ea1f5080626b name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.261549647Z" level=info msg="Started container" PID=6814 containerID=2c1936a19655345a892373f86a623f7409e46829cbb1adeca75e2879d268aa08 description=openshift-monitoring/telemeter-client-5598c645c4-mftdv/telemeter-client id=5fa8b7ff-e67e-4f11-88a3-58e85f85f349 name=/runtime.v1.RuntimeService/StartContainer sandboxID=f9cc29d007bd3c948c46f781abd08afdb46359ec50fc31f8c503fcdb35b7aa22 Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.270353764Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7" id=0cbff82e-d919-4855-b88c-953fd5c341d2 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.270599706Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:4cb41c73a232ec4401c5db32cdb523810dc12ac605cdc6a9ce388e90930eb6c4,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7],Size_:515955107,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=0cbff82e-d919-4855-b88c-953fd5c341d2 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.301638450Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7" id=e17cb743-fb57-41db-a618-209a9a5ad99f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.301868594Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:4cb41c73a232ec4401c5db32cdb523810dc12ac605cdc6a9ce388e90930eb6c4,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7],Size_:515955107,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=e17cb743-fb57-41db-a618-209a9a5ad99f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.303545975Z" level=info msg="Creating container: openshift-network-diagnostics/network-check-target-xmq2g/network-check-target-container" id=ef29c36c-f25e-417b-8f37-070e2645075d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.303689069Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:13.327306 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-2pfzf" event=&{ID:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 Type:ContainerStarted Data:4f0a921bd0d3c8f309751665c310e818034ecd3252406f2a101ca811d04377c5} Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:13.330061 2588 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256] Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:13.340606 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" event=&{ID:7c63c9be-d1ad-44b6-8b33-09af3ee314af Type:ContainerStarted Data:578cce0ed1aa4795c3b6d8648ecf4b937233d871e6a027eb233bec20d4333c1f} Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:13.341844 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" event=&{ID:fc45e847-8f55-4c48-8c48-4bfee2419dca Type:ContainerStarted Data:2c1936a19655345a892373f86a623f7409e46829cbb1adeca75e2879d268aa08} Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:13.341876 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" event=&{ID:fc45e847-8f55-4c48-8c48-4bfee2419dca Type:ContainerStarted Data:f9cc29d007bd3c948c46f781abd08afdb46359ec50fc31f8c503fcdb35b7aa22} Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:13.344156 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event=&{ID:05c9a6ea-140a-49a2-bde5-f220b05aa252 Type:ContainerStarted Data:4c970eddded2809e2e506cfb0b207a19b375bc49b3a1f9bfd58b202b7c1f6ebd} Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:13.344240 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event=&{ID:05c9a6ea-140a-49a2-bde5-f220b05aa252 Type:ContainerStarted Data:53a26d2c619d3ef8230bda385bc9245dc43ada80f94fe54f1f79b82763d2f7a8} Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:13.345967 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xmq2g" event=&{ID:f364a949-09d7-4cfb-83ff-e532b822a557 Type:ContainerStarted Data:df03e4f874ec90019c812fc416c2d12340d64e387a2a542bdc3beee879bc31bd} Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:42:13Z [verbose] Add: openshift-monitoring:openshift-state-metrics-6f5dcdf79b-wq256:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/e7429438-8b02-46d6-a53c-f91c31850d37"}],"ips":[{"version":"4","interface":0,"address":"10.131.0.11/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: I0105 09:42:13.275671 6543 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"openshift-state-metrics-6f5dcdf79b-wq256", UID:"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6", APIVersion:"v1", ResourceVersion:"96790", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.11/23] from openshift-sdn Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.388566434Z" level=info msg="Got pod network &{Name:openshift-state-metrics-6f5dcdf79b-wq256 Namespace:openshift-monitoring ID:7a37710cfebf59892a4d6b3c2e12765b60414f6a7cfb8933a050df609c94e690 UID:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 NetNS:/var/run/netns/e7429438-8b02-46d6-a53c-f91c31850d37 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.388712800Z" level=info msg="Checking pod openshift-monitoring_openshift-state-metrics-6f5dcdf79b-wq256 for CNI network multus-cni-network (type=multus)" Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.407783455Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=da5707a4-584c-4ee6-bfe7-82bb974855cd name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:13.418416 2588 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.450521989Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=da5707a4-584c-4ee6-bfe7-82bb974855cd name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.465607105Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=59153fef-6a4a-436b-98d9-de9db08c9e13 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.465805189Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=59153fef-6a4a-436b-98d9-de9db08c9e13 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.466751537Z" level=info msg="Creating container: openshift-monitoring/telemeter-client-5598c645c4-mftdv/reload" id=47f63eed-f39f-4d7d-a114-28fbaa6d1c16 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.466862355Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:42:13.466534 2588 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9de5de36_9b80_4a0b_a615_f0f9ccda2ff6.slice/crio-7a37710cfebf59892a4d6b3c2e12765b60414f6a7cfb8933a050df609c94e690.scope WatchSource:0}: Error finding container 7a37710cfebf59892a4d6b3c2e12765b60414f6a7cfb8933a050df609c94e690: Status 404 returned error can't find the container with id 7a37710cfebf59892a4d6b3c2e12765b60414f6a7cfb8933a050df609c94e690 Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:13.481799 2588 patch_prober.go:29] interesting pod/router-default-5cf6bbdff9-tnffs container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: [-]has-synced failed: reason withheld Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: [+]process-running ok Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: healthz check failed Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:13.481853 2588 prober.go:114] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" podUID=c3b10ff4-a243-4a4e-b27d-f613e258666d containerName="router" probeResult=failure output="HTTP probe failed with statuscode: 500" Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.482036749Z" level=info msg="Created container 76d0bbfe3b04b3fb31c951bf14c59dddccfd473352dd9388eeee0e7e51e96358: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/oauth-proxy" id=8becde2f-45fb-41c9-9f31-695820ba74d3 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.487322113Z" level=info msg="Ran pod sandbox 7a37710cfebf59892a4d6b3c2e12765b60414f6a7cfb8933a050df609c94e690 with infra container: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/POD" id=fbf40af3-df42-4b07-a1a3-974b71a0b0e1 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.488143621Z" level=info msg="Starting container: 76d0bbfe3b04b3fb31c951bf14c59dddccfd473352dd9388eeee0e7e51e96358" id=622190ca-bc4a-4df9-a492-8650c75696a4 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.493606898Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=9757e151-1270-465a-b07a-4ab5545b44fd name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.503131411Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=9757e151-1270-465a-b07a-4ab5545b44fd name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.516589650Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=4c3c7819-5cfe-4dd0-9469-70208c29b6bf name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.516788315Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=4c3c7819-5cfe-4dd0-9469-70208c29b6bf name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.517719773Z" level=info msg="Creating container: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/kube-rbac-proxy-main" id=e9dde0fd-4e4c-49bd-b480-713bcaebe305 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.517824809Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.567393311Z" level=info msg="Started container" PID=6891 containerID=76d0bbfe3b04b3fb31c951bf14c59dddccfd473352dd9388eeee0e7e51e96358 description=openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/oauth-proxy id=622190ca-bc4a-4df9-a492-8650c75696a4 name=/runtime.v1.RuntimeService/StartContainer sandboxID=fcdbc115e8986f7ba4e1fc9bfa0728fd149a52733e5f0956b91d0b8262b33ca6 Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.694409276Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=77fc67d2-fa46-47e1-892c-00a6c8dce7d3 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.694658325Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=77fc67d2-fa46-47e1-892c-00a6c8dce7d3 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.696909402Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=c8e67794-fd94-4f9a-afaf-68929f9c11a7 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.697116145Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=c8e67794-fd94-4f9a-afaf-68929f9c11a7 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.698246998Z" level=info msg="Creating container: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy" id=d212fae1-ac68-46cf-bbcd-c7d5b43fc8d8 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.698386757Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.782293306Z" level=info msg="Created container 4a95d41273862399b3cca16471b999defcccef3ceb8402d7cfb798378ffc8cab: openshift-monitoring/alertmanager-main-0/prom-label-proxy" id=fb4c36ac-0c9d-47ce-a14d-6c5c16f7aa3e name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:13.791689 2588 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/kube-state-metrics-6798b548b9-xl88q] Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.804067197Z" level=info msg="Starting container: 4a95d41273862399b3cca16471b999defcccef3ceb8402d7cfb798378ffc8cab" id=9b0858ff-7e7d-4de5-8897-d6fe43fefb38 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:42:13Z [verbose] Add: openshift-monitoring:kube-state-metrics-6798b548b9-xl88q:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/cbe37ad3-4a92-423a-8f3c-dda00e586a03"}],"ips":[{"version":"4","interface":0,"address":"10.131.0.12/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: I0105 09:42:13.704475 6552 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"kube-state-metrics-6798b548b9-xl88q", UID:"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e", APIVersion:"v1", ResourceVersion:"96807", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.12/23] from openshift-sdn Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.807431237Z" level=info msg="Got pod network &{Name:kube-state-metrics-6798b548b9-xl88q Namespace:openshift-monitoring ID:1e97b2e64ee6d4f76d829e9619da041f24e571e3de5454c48314d411b8770c02 UID:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e NetNS:/var/run/netns/cbe37ad3-4a92-423a-8f3c-dda00e586a03 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.807575825Z" level=info msg="Checking pod openshift-monitoring_kube-state-metrics-6798b548b9-xl88q for CNI network multus-cni-network (type=multus)" Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:42:13.814080 2588 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ecd2e2c_ab52_42e0_8fde_64e45804cc9e.slice/crio-1e97b2e64ee6d4f76d829e9619da041f24e571e3de5454c48314d411b8770c02.scope WatchSource:0}: Error finding container 1e97b2e64ee6d4f76d829e9619da041f24e571e3de5454c48314d411b8770c02: Status 404 returned error can't find the container with id 1e97b2e64ee6d4f76d829e9619da041f24e571e3de5454c48314d411b8770c02 Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.833470691Z" level=info msg="Ran pod sandbox 1e97b2e64ee6d4f76d829e9619da041f24e571e3de5454c48314d411b8770c02 with infra container: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/POD" id=e91c6a60-ff52-4909-a60c-3d6c50f02a0c name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.838759188Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e5521a3d1f64483e63ffe5bc23d9f882781cd4d152ac6de72013eaefe6dfcf89" id=b05c4a0c-ce48-453d-84a7-9d372cd4b873 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.838972076Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:af30ce06a0837328332ff57ed27c48095a8ed9600535c919aa69fa76524919c6,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e5521a3d1f64483e63ffe5bc23d9f882781cd4d152ac6de72013eaefe6dfcf89],Size_:371510113,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b05c4a0c-ce48-453d-84a7-9d372cd4b873 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.865399088Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e5521a3d1f64483e63ffe5bc23d9f882781cd4d152ac6de72013eaefe6dfcf89" id=80d0ddfc-f352-45a0-a3cb-0d9b6fc0166b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.865608411Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:af30ce06a0837328332ff57ed27c48095a8ed9600535c919aa69fa76524919c6,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e5521a3d1f64483e63ffe5bc23d9f882781cd4d152ac6de72013eaefe6dfcf89],Size_:371510113,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=80d0ddfc-f352-45a0-a3cb-0d9b6fc0166b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.866792091Z" level=info msg="Creating container: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-state-metrics" id=91af60e0-2eab-41a7-9ccb-4e41845a7ac6 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.866902282Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.892433153Z" level=info msg="Started container" PID=7012 containerID=4a95d41273862399b3cca16471b999defcccef3ceb8402d7cfb798378ffc8cab description=openshift-monitoring/alertmanager-main-0/prom-label-proxy id=9b0858ff-7e7d-4de5-8897-d6fe43fefb38 name=/runtime.v1.RuntimeService/StartContainer sandboxID=d6d0c82bc9774e51559f70121d26a57c7c38cce6fb83f23b63a8bb671c408015 Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.899222392Z" level=info msg="Created container 4f1c513a2387ed2e4cba878f4c591e83420e71f65da912f971c12aa0fc9e8d05: openshift-network-diagnostics/network-check-target-xmq2g/network-check-target-container" id=ef29c36c-f25e-417b-8f37-070e2645075d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.899632008Z" level=info msg="Starting container: 4f1c513a2387ed2e4cba878f4c591e83420e71f65da912f971c12aa0fc9e8d05" id=c5bfd3ef-c703-452d-b0a8-2fdc25d381c4 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.972877958Z" level=info msg="Started container" PID=7041 containerID=4f1c513a2387ed2e4cba878f4c591e83420e71f65da912f971c12aa0fc9e8d05 description=openshift-network-diagnostics/network-check-target-xmq2g/network-check-target-container id=c5bfd3ef-c703-452d-b0a8-2fdc25d381c4 name=/runtime.v1.RuntimeService/StartContainer sandboxID=df03e4f874ec90019c812fc416c2d12340d64e387a2a542bdc3beee879bc31bd Jan 05 09:42:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.995868650Z" level=info msg="Created container 2952a168f0605321dd4ff798bf4481d0ede73526e359849e9e7039226b853fab: openshift-monitoring/telemeter-client-5598c645c4-mftdv/reload" id=47f63eed-f39f-4d7d-a114-28fbaa6d1c16 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:13.996957610Z" level=info msg="Starting container: 2952a168f0605321dd4ff798bf4481d0ede73526e359849e9e7039226b853fab" id=c3eec1ce-084f-4037-8452-b1fed508462b name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.024107255Z" level=info msg="Started container" PID=7053 containerID=2952a168f0605321dd4ff798bf4481d0ede73526e359849e9e7039226b853fab description=openshift-monitoring/telemeter-client-5598c645c4-mftdv/reload id=c3eec1ce-084f-4037-8452-b1fed508462b name=/runtime.v1.RuntimeService/StartContainer sandboxID=f9cc29d007bd3c948c46f781abd08afdb46359ec50fc31f8c503fcdb35b7aa22 Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.083081431Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=2da71c39-2e81-474b-ab4b-2556d2e74fc4 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.083340834Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=2da71c39-2e81-474b-ab4b-2556d2e74fc4 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.097478160Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=9d4eb6d0-8121-4214-a813-e335c11b4a12 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.097699521Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=9d4eb6d0-8121-4214-a813-e335c11b4a12 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.099015138Z" level=info msg="Creating container: openshift-monitoring/telemeter-client-5598c645c4-mftdv/kube-rbac-proxy" id=9f70d14e-f0e7-40dd-8329-226dd0b36ab5 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.099134264Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.103260920Z" level=info msg="Created container 10f645c424bd560dd410724c8e884a10049cfe720e59016dcb3f8d51623b3ae6: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/kube-rbac-proxy-main" id=e9dde0fd-4e4c-49bd-b480-713bcaebe305 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.112254444Z" level=info msg="Starting container: 10f645c424bd560dd410724c8e884a10049cfe720e59016dcb3f8d51623b3ae6" id=d6a0caa8-8675-4367-80f4-d8f7f21e4f07 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.131634284Z" level=info msg="NetworkStart: stopping network for sandbox a108e4a07ee46c33fed5d4508ec6bf5478eacc0dc17a5ac7b7e9bfbd9fa620d3" id=71e3a947-76ff-4d18-a8dc-205bbf21124d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.131895611Z" level=info msg="Got pod network &{Name:prometheus-k8s-0 Namespace:openshift-monitoring ID:a108e4a07ee46c33fed5d4508ec6bf5478eacc0dc17a5ac7b7e9bfbd9fa620d3 UID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 NetNS:/var/run/netns/7b579e74-f3d2-4d6d-990a-eb17dda53649 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.131936407Z" level=error msg="error loading cached network config: network \"multus-cni-network\" not found in CNI cache" Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.131947914Z" level=warning msg="falling back to loading from existing plugins on disk" Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.131958511Z" level=info msg="Deleting pod openshift-monitoring_prometheus-k8s-0 from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.137258195Z" level=info msg="Started container" PID=7058 containerID=10f645c424bd560dd410724c8e884a10049cfe720e59016dcb3f8d51623b3ae6 description=openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/kube-rbac-proxy-main id=d6a0caa8-8675-4367-80f4-d8f7f21e4f07 name=/runtime.v1.RuntimeService/StartContainer sandboxID=7a37710cfebf59892a4d6b3c2e12765b60414f6a7cfb8933a050df609c94e690 Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.242568872Z" level=info msg="Created container aee14ddf06d77378c3afc5489f1ac566eb37f4e3542ad949a8b18afeff164f94: openshift-dns/dns-default-2pfzf/dns" id=8386fadb-94fb-4fa1-919b-71a124e20e93 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.243126556Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=fd5844b2-d825-4ce6-90f2-9b7f8d494df5 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.243373076Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=fd5844b2-d825-4ce6-90f2-9b7f8d494df5 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.243498023Z" level=info msg="Starting container: aee14ddf06d77378c3afc5489f1ac566eb37f4e3542ad949a8b18afeff164f94" id=be5fa7e0-55c2-4bf0-b388-806118a094b5 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.246839184Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=552334e3-aa38-4969-94af-7e8885b7198f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.247081045Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=552334e3-aa38-4969-94af-7e8885b7198f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.254809205Z" level=info msg="Creating container: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/kube-rbac-proxy-self" id=e155576b-d3b9-4f00-804f-341f0c6a7713 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.254938289Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.292149563Z" level=info msg="Started container" PID=7160 containerID=aee14ddf06d77378c3afc5489f1ac566eb37f4e3542ad949a8b18afeff164f94 description=openshift-dns/dns-default-2pfzf/dns id=be5fa7e0-55c2-4bf0-b388-806118a094b5 name=/runtime.v1.RuntimeService/StartContainer sandboxID=4f0a921bd0d3c8f309751665c310e818034ecd3252406f2a101ca811d04377c5 Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.341087729Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=edb209ba-1350-48aa-8506-255f69d36b2f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.341361217Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=edb209ba-1350-48aa-8506-255f69d36b2f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.344717558Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=16789970-af95-4645-92d3-50a1c2a39725 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.344913490Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=16789970-af95-4645-92d3-50a1c2a39725 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.346249274Z" level=info msg="Creating container: openshift-dns/dns-default-2pfzf/kube-rbac-proxy" id=95f29516-148a-4a91-85fe-607782e4e669 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.346368867Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:14.360046 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" event=&{ID:7c63c9be-d1ad-44b6-8b33-09af3ee314af Type:ContainerStarted Data:76d0bbfe3b04b3fb31c951bf14c59dddccfd473352dd9388eeee0e7e51e96358} Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:14.361871 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" event=&{ID:fc45e847-8f55-4c48-8c48-4bfee2419dca Type:ContainerStarted Data:2952a168f0605321dd4ff798bf4481d0ede73526e359849e9e7039226b853fab} Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:14.363036 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-2pfzf" event=&{ID:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 Type:ContainerStarted Data:aee14ddf06d77378c3afc5489f1ac566eb37f4e3542ad949a8b18afeff164f94} Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:14.364166 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" event=&{ID:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 Type:ContainerStarted Data:10f645c424bd560dd410724c8e884a10049cfe720e59016dcb3f8d51623b3ae6} Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:14.364216 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" event=&{ID:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 Type:ContainerStarted Data:7a37710cfebf59892a4d6b3c2e12765b60414f6a7cfb8933a050df609c94e690} Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:14.366969 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event=&{ID:05c9a6ea-140a-49a2-bde5-f220b05aa252 Type:ContainerStarted Data:4a95d41273862399b3cca16471b999defcccef3ceb8402d7cfb798378ffc8cab} Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:14.371314 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xmq2g" event=&{ID:f364a949-09d7-4cfb-83ff-e532b822a557 Type:ContainerStarted Data:4f1c513a2387ed2e4cba878f4c591e83420e71f65da912f971c12aa0fc9e8d05} Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:14.371952 2588 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:42:14Z [verbose] Del: openshift-monitoring:prometheus-k8s-0:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.375881022Z" level=info msg="Got pod network &{Name:prometheus-k8s-0 Namespace:openshift-monitoring ID:04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0 UID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.376056598Z" level=info msg="Deleting pod openshift-monitoring_prometheus-k8s-0 from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:14.375900 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" event=&{ID:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e Type:ContainerStarted Data:1e97b2e64ee6d4f76d829e9619da041f24e571e3de5454c48314d411b8770c02} Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.376701315Z" level=info msg="runSandbox: cleaning up namespaces after failing to run sandbox a108e4a07ee46c33fed5d4508ec6bf5478eacc0dc17a5ac7b7e9bfbd9fa620d3" id=71e3a947-76ff-4d18-a8dc-205bbf21124d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.391618718Z" level=info msg="Created container c507c8393ab5e7e8bb6dbc97bcabc3f2598889f9b13fbee87614f2fee17b1e9d: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy" id=d212fae1-ac68-46cf-bbcd-c7d5b43fc8d8 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.401079502Z" level=info msg="Starting container: c507c8393ab5e7e8bb6dbc97bcabc3f2598889f9b13fbee87614f2fee17b1e9d" id=c89dceec-65c9-4d06-b4a0-eb78a0a562ea name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.439527885Z" level=info msg="Started container" PID=7184 containerID=c507c8393ab5e7e8bb6dbc97bcabc3f2598889f9b13fbee87614f2fee17b1e9d description=openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy id=c89dceec-65c9-4d06-b4a0-eb78a0a562ea name=/runtime.v1.RuntimeService/StartContainer sandboxID=fcdbc115e8986f7ba4e1fc9bfa0728fd149a52733e5f0956b91d0b8262b33ca6 Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:14.450806 2588 patch_prober.go:29] interesting pod/router-default-5cf6bbdff9-tnffs container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: [-]has-synced failed: reason withheld Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: [+]process-running ok Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: healthz check failed Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:14.450867 2588 prober.go:114] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" podUID=c3b10ff4-a243-4a4e-b27d-f613e258666d containerName="router" probeResult=failure output="HTTP probe failed with statuscode: 500" Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.511502849Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816" id=3403ad98-8fda-4c5a-9f58-75e64ae1766f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.511754156Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:955ee2b41ecd88e3a9c0183a0b65d6edbf524cefa1318837dcbc82b0466c4d55,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816],Size_:309576938,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=3403ad98-8fda-4c5a-9f58-75e64ae1766f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.522633832Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816" id=c897a6f2-142d-423a-adbf-1b5dcddd3707 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.522878198Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:955ee2b41ecd88e3a9c0183a0b65d6edbf524cefa1318837dcbc82b0466c4d55,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816],Size_:309576938,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=c897a6f2-142d-423a-adbf-1b5dcddd3707 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.546143948Z" level=info msg="runSandbox: deleting pod ID a108e4a07ee46c33fed5d4508ec6bf5478eacc0dc17a5ac7b7e9bfbd9fa620d3 from idIndex" id=71e3a947-76ff-4d18-a8dc-205bbf21124d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.546207843Z" level=info msg="runSandbox: removing pod sandbox a108e4a07ee46c33fed5d4508ec6bf5478eacc0dc17a5ac7b7e9bfbd9fa620d3" id=71e3a947-76ff-4d18-a8dc-205bbf21124d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.546231018Z" level=info msg="runSandbox: deleting container ID from idIndex for sandbox a108e4a07ee46c33fed5d4508ec6bf5478eacc0dc17a5ac7b7e9bfbd9fa620d3" id=71e3a947-76ff-4d18-a8dc-205bbf21124d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.546254513Z" level=info msg="runSandbox: unmounting shmPath for sandbox a108e4a07ee46c33fed5d4508ec6bf5478eacc0dc17a5ac7b7e9bfbd9fa620d3" id=71e3a947-76ff-4d18-a8dc-205bbf21124d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.560359093Z" level=info msg="Created container 987ad82f0f8b3760e232b876a53eb03f7fa455eadddd9e240e5e8be4f57bf369: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-state-metrics" id=91af60e0-2eab-41a7-9ccb-4e41845a7ac6 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.569018269Z" level=info msg="Creating container: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/prom-label-proxy" id=c219ed80-5f08-436d-aaec-df750af7f15d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.569148508Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.634611649Z" level=info msg="runSandbox: removing pod sandbox from storage: a108e4a07ee46c33fed5d4508ec6bf5478eacc0dc17a5ac7b7e9bfbd9fa620d3" id=71e3a947-76ff-4d18-a8dc-205bbf21124d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.664171486Z" level=info msg="Starting container: 987ad82f0f8b3760e232b876a53eb03f7fa455eadddd9e240e5e8be4f57bf369" id=a183365f-b20a-4e56-a769-b790fbb73097 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.749067552Z" level=info msg="runSandbox: releasing container name: k8s_POD_prometheus-k8s-0_openshift-monitoring_4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879_0" id=71e3a947-76ff-4d18-a8dc-205bbf21124d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.749107701Z" level=info msg="runSandbox: releasing pod sandbox name: k8s_prometheus-k8s-0_openshift-monitoring_4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879_0" id=71e3a947-76ff-4d18-a8dc-205bbf21124d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:42:14.757816 2588 remote_runtime.go:222] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_prometheus-k8s-0_openshift-monitoring_4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879_0(a108e4a07ee46c33fed5d4508ec6bf5478eacc0dc17a5ac7b7e9bfbd9fa620d3): error adding pod openshift-monitoring_prometheus-k8s-0 to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/prometheus-k8s-0/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:42:14.757902 2588 kuberuntime_sandbox.go:71] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_prometheus-k8s-0_openshift-monitoring_4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879_0(a108e4a07ee46c33fed5d4508ec6bf5478eacc0dc17a5ac7b7e9bfbd9fa620d3): error adding pod openshift-monitoring_prometheus-k8s-0 to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/prometheus-k8s-0/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:42:14.757945 2588 kuberuntime_manager.go:772] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_prometheus-k8s-0_openshift-monitoring_4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879_0(a108e4a07ee46c33fed5d4508ec6bf5478eacc0dc17a5ac7b7e9bfbd9fa620d3): error adding pod openshift-monitoring_prometheus-k8s-0 to CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (add): Multus: [openshift-monitoring/prometheus-k8s-0/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition" pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:42:14.758039 2588 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"prometheus-k8s-0_openshift-monitoring(4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"prometheus-k8s-0_openshift-monitoring(4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_prometheus-k8s-0_openshift-monitoring_4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879_0(a108e4a07ee46c33fed5d4508ec6bf5478eacc0dc17a5ac7b7e9bfbd9fa620d3): error adding pod openshift-monitoring_prometheus-k8s-0 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus\\\" name=\\\"multus-cni-network\\\" failed (add): Multus: [openshift-monitoring/prometheus-k8s-0/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879]: have you checked that your default network is ready? still waiting for readinessindicatorfile @ /var/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\"" pod="openshift-monitoring/prometheus-k8s-0" podUID=4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.764829171Z" level=info msg="Created container 91947bee1cff242d8ff480a4ba98a9c65dd41326ef2bfc74ccc1a969906e3b00: openshift-monitoring/telemeter-client-5598c645c4-mftdv/kube-rbac-proxy" id=9f70d14e-f0e7-40dd-8329-226dd0b36ab5 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.791009241Z" level=info msg="Started container" PID=7186 containerID=987ad82f0f8b3760e232b876a53eb03f7fa455eadddd9e240e5e8be4f57bf369 description=openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-state-metrics id=a183365f-b20a-4e56-a769-b790fbb73097 name=/runtime.v1.RuntimeService/StartContainer sandboxID=1e97b2e64ee6d4f76d829e9619da041f24e571e3de5454c48314d411b8770c02 Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.805534745Z" level=info msg="Starting container: 91947bee1cff242d8ff480a4ba98a9c65dd41326ef2bfc74ccc1a969906e3b00" id=e9348114-f801-4129-97a3-68c85aae0404 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.881398793Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=2fa46e26-2716-4121-bbe0-d45c7d68978f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.881631503Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=2fa46e26-2716-4121-bbe0-d45c7d68978f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.881817318Z" level=info msg="Started container" PID=7241 containerID=91947bee1cff242d8ff480a4ba98a9c65dd41326ef2bfc74ccc1a969906e3b00 description=openshift-monitoring/telemeter-client-5598c645c4-mftdv/kube-rbac-proxy id=e9348114-f801-4129-97a3-68c85aae0404 name=/runtime.v1.RuntimeService/StartContainer sandboxID=f9cc29d007bd3c948c46f781abd08afdb46359ec50fc31f8c503fcdb35b7aa22 Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.892773240Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=0f53b1d0-2e25-4248-8853-a7a9e8348928 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.892984217Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=0f53b1d0-2e25-4248-8853-a7a9e8348928 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.894208507Z" level=info msg="Creating container: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-rbac-proxy-main" id=4e098f8b-a52f-4c60-b823-ee47203e760d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:14.894334000Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.147812855Z" level=info msg="Created container 489142621cb0989124915d3c9efa7b65c6ff3af9ed775b03b9f887e6ad7d1d8b: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/kube-rbac-proxy-self" id=e155576b-d3b9-4f00-804f-341f0c6a7713 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.155419504Z" level=info msg="Starting container: 489142621cb0989124915d3c9efa7b65c6ff3af9ed775b03b9f887e6ad7d1d8b" id=dc09cff9-0b2d-4f1c-a22c-9eef3c31d17c name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.170035645Z" level=info msg="Created container 050203bff5fd4bea33b4472ea6e147349a7124fa32c21b73f71e99e34b94276d: openshift-dns/dns-default-2pfzf/kube-rbac-proxy" id=95f29516-148a-4a91-85fe-607782e4e669 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.171638560Z" level=info msg="Starting container: 050203bff5fd4bea33b4472ea6e147349a7124fa32c21b73f71e99e34b94276d" id=0d99f705-6b6d-420f-b32a-527ce4f15dfb name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.183625772Z" level=info msg="Started container" PID=7320 containerID=489142621cb0989124915d3c9efa7b65c6ff3af9ed775b03b9f887e6ad7d1d8b description=openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/kube-rbac-proxy-self id=dc09cff9-0b2d-4f1c-a22c-9eef3c31d17c name=/runtime.v1.RuntimeService/StartContainer sandboxID=7a37710cfebf59892a4d6b3c2e12765b60414f6a7cfb8933a050df609c94e690 Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.257911402Z" level=info msg="Started container" PID=7396 containerID=050203bff5fd4bea33b4472ea6e147349a7124fa32c21b73f71e99e34b94276d description=openshift-dns/dns-default-2pfzf/kube-rbac-proxy id=0d99f705-6b6d-420f-b32a-527ce4f15dfb name=/runtime.v1.RuntimeService/StartContainer sandboxID=4f0a921bd0d3c8f309751665c310e818034ecd3252406f2a101ca811d04377c5 Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.292525059Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:78252776c28ff0ef5195173aebd633dc3a71e5d4732ca01f9104bb0aab07570e" id=ff36e2cf-902d-4b6b-88b5-8af018086ff2 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.292791095Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:ab19b64ab85c6b98c0381b09c3bd6f6db37bb31b25f12aa07bf856e8be07eac7,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:78252776c28ff0ef5195173aebd633dc3a71e5d4732ca01f9104bb0aab07570e],Size_:365012751,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=ff36e2cf-902d-4b6b-88b5-8af018086ff2 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.295837566Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:78252776c28ff0ef5195173aebd633dc3a71e5d4732ca01f9104bb0aab07570e" id=9bdceba2-16f1-472d-95ec-0b5c6349d23d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.296056885Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:ab19b64ab85c6b98c0381b09c3bd6f6db37bb31b25f12aa07bf856e8be07eac7,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:78252776c28ff0ef5195173aebd633dc3a71e5d4732ca01f9104bb0aab07570e],Size_:365012751,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=9bdceba2-16f1-472d-95ec-0b5c6349d23d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.297330298Z" level=info msg="Creating container: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/openshift-state-metrics" id=77e9ef37-f784-4217-a58b-f632bcd7df9e name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.297468588Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.313563541Z" level=info msg="Created container b7574d11ea7895beed8a146040a8adb91835493e426424dd8e461a93b6f7677b: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/prom-label-proxy" id=c219ed80-5f08-436d-aaec-df750af7f15d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.318720266Z" level=info msg="Starting container: b7574d11ea7895beed8a146040a8adb91835493e426424dd8e461a93b6f7677b" id=01e7f5ed-6bd3-435f-86ec-ab6614344f80 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.339717992Z" level=info msg="Started container" PID=7384 containerID=b7574d11ea7895beed8a146040a8adb91835493e426424dd8e461a93b6f7677b description=openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/prom-label-proxy id=01e7f5ed-6bd3-435f-86ec-ab6614344f80 name=/runtime.v1.RuntimeService/StartContainer sandboxID=fcdbc115e8986f7ba4e1fc9bfa0728fd149a52733e5f0956b91d0b8262b33ca6 Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.392962986Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=e5d3e724-f058-42c0-9fce-79ce8463ca62 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.393322699Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=e5d3e724-f058-42c0-9fce-79ce8463ca62 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.407921730Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=3b1b6c65-84bd-4635-9687-000fe35b902c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.408134367Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=3b1b6c65-84bd-4635-9687-000fe35b902c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.408876605Z" level=info msg="Creating container: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy-rules" id=cf9e9118-7402-40b9-bef4-7ef51631a607 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.408992411Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:15.427339 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-2pfzf" event=&{ID:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 Type:ContainerStarted Data:050203bff5fd4bea33b4472ea6e147349a7124fa32c21b73f71e99e34b94276d} Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:15.428407 2588 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-2pfzf" Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:42:15Z [verbose] Add: openshift-monitoring:prometheus-adapter-5894cc86c7-qhmbh:a40bf20d-67a4-45e7-b701-ca5a75376c1f:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/2bb8265e-9d3b-4130-ab12-f8f63e286c5d"}],"ips":[{"version":"4","interface":0,"address":"10.131.0.13/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: I0105 09:42:15.295693 6844 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"prometheus-adapter-5894cc86c7-qhmbh", UID:"a40bf20d-67a4-45e7-b701-ca5a75376c1f", APIVersion:"v1", ResourceVersion:"96798", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.13/23] from openshift-sdn Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.447437356Z" level=info msg="Got pod network &{Name:prometheus-adapter-5894cc86c7-qhmbh Namespace:openshift-monitoring ID:9d9807a6998df2507a4aa5b9bad79e347bc60d62b254989738c927929d827e6e UID:a40bf20d-67a4-45e7-b701-ca5a75376c1f NetNS:/var/run/netns/2bb8265e-9d3b-4130-ab12-f8f63e286c5d Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.447583393Z" level=info msg="Checking pod openshift-monitoring_prometheus-adapter-5894cc86c7-qhmbh for CNI network multus-cni-network (type=multus)" Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:15.449134 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" event=&{ID:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 Type:ContainerStarted Data:489142621cb0989124915d3c9efa7b65c6ff3af9ed775b03b9f887e6ad7d1d8b} Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:15.452891 2588 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh] Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:15.455770 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" event=&{ID:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e Type:ContainerStarted Data:987ad82f0f8b3760e232b876a53eb03f7fa455eadddd9e240e5e8be4f57bf369} Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.458707174Z" level=info msg="Ran pod sandbox 9d9807a6998df2507a4aa5b9bad79e347bc60d62b254989738c927929d827e6e with infra container: openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh/POD" id=a63a286d-5166-431d-a81d-ecfec74c8af7 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.467208698Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9696f8f231d49aadf63ab980ec5c17c4fd87011930daecd30cc9e4ee102f1aeb" id=7d1e9019-07cc-4ce8-ace9-8f909deb97b2 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.467571084Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:ea0b364e3a8a4e4b3f4ab3288e42c5be8021187d32791029f9cb9a500c8da100,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9696f8f231d49aadf63ab980ec5c17c4fd87011930daecd30cc9e4ee102f1aeb],Size_:402351209,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=7d1e9019-07cc-4ce8-ace9-8f909deb97b2 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:15.492110 2588 kubelet.go:2229] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.499458121Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9696f8f231d49aadf63ab980ec5c17c4fd87011930daecd30cc9e4ee102f1aeb" id=36ecf88a-2930-45d9-a488-c574399af942 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.499682256Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:ea0b364e3a8a4e4b3f4ab3288e42c5be8021187d32791029f9cb9a500c8da100,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9696f8f231d49aadf63ab980ec5c17c4fd87011930daecd30cc9e4ee102f1aeb],Size_:402351209,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=36ecf88a-2930-45d9-a488-c574399af942 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.501095274Z" level=info msg="Creating container: openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh/prometheus-adapter" id=6e2cbb62-3354-4a07-9f3b-64777e2272c5 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.501229408Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:15.516345 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" event=&{ID:7c63c9be-d1ad-44b6-8b33-09af3ee314af Type:ContainerStarted Data:b7574d11ea7895beed8a146040a8adb91835493e426424dd8e461a93b6f7677b} Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:15.516401 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" event=&{ID:7c63c9be-d1ad-44b6-8b33-09af3ee314af Type:ContainerStarted Data:c507c8393ab5e7e8bb6dbc97bcabc3f2598889f9b13fbee87614f2fee17b1e9d} Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:15.532438 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" event=&{ID:fc45e847-8f55-4c48-8c48-4bfee2419dca Type:ContainerStarted Data:91947bee1cff242d8ff480a4ba98a9c65dd41326ef2bfc74ccc1a969906e3b00} Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:15.561512 2588 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.584643860Z" level=info msg="Created container 46e4da5180913795c57fd23e927474161e2562ce8f557a4ef7d187e1da8df07c: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-rbac-proxy-main" id=4e098f8b-a52f-4c60-b823-ee47203e760d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.586143327Z" level=info msg="Starting container: 46e4da5180913795c57fd23e927474161e2562ce8f557a4ef7d187e1da8df07c" id=c237c728-0dbe-4966-99a5-3abed19dc4f3 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.667034108Z" level=info msg="Started container" PID=7482 containerID=46e4da5180913795c57fd23e927474161e2562ce8f557a4ef7d187e1da8df07c description=openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-rbac-proxy-main id=c237c728-0dbe-4966-99a5-3abed19dc4f3 name=/runtime.v1.RuntimeService/StartContainer sandboxID=1e97b2e64ee6d4f76d829e9619da041f24e571e3de5454c48314d411b8770c02 Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.722704756Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=699cc266-5074-4fdc-8d8d-3330222d06dd name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.722950569Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=699cc266-5074-4fdc-8d8d-3330222d06dd name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.724329919Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=166691c9-c958-4b9b-8848-e6a013f8e83e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.724533055Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=166691c9-c958-4b9b-8848-e6a013f8e83e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.728099628Z" level=info msg="Creating container: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-rbac-proxy-self" id=31a02fdf-191d-427c-8f03-8b79907576f3 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.728240513Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.969896230Z" level=info msg="Created container 0f22074c4c974c98091736b511fa5de52c01939cfdd9697e4eee429370ad1540: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/openshift-state-metrics" id=77e9ef37-f784-4217-a58b-f632bcd7df9e name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.970762311Z" level=info msg="Starting container: 0f22074c4c974c98091736b511fa5de52c01939cfdd9697e4eee429370ad1540" id=7c2678e1-0c3b-47bd-9c9b-acad0e9da277 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.975562632Z" level=info msg="Created container 3cbc4c089943552285b7235ee6268116978902d078a993e4f88285e0c16777ce: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy-rules" id=cf9e9118-7402-40b9-bef4-7ef51631a607 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:15.977239693Z" level=info msg="Starting container: 3cbc4c089943552285b7235ee6268116978902d078a993e4f88285e0c16777ce" id=711320ca-9677-4017-ab5e-dcc9ceb4d7e8 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:16.031760141Z" level=info msg="Started container" PID=7597 containerID=3cbc4c089943552285b7235ee6268116978902d078a993e4f88285e0c16777ce description=openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy-rules id=711320ca-9677-4017-ab5e-dcc9ceb4d7e8 name=/runtime.v1.RuntimeService/StartContainer sandboxID=fcdbc115e8986f7ba4e1fc9bfa0728fd149a52733e5f0956b91d0b8262b33ca6 Jan 05 09:42:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:16.031977197Z" level=info msg="Started container" PID=7567 containerID=0f22074c4c974c98091736b511fa5de52c01939cfdd9697e4eee429370ad1540 description=openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/openshift-state-metrics id=7c2678e1-0c3b-47bd-9c9b-acad0e9da277 name=/runtime.v1.RuntimeService/StartContainer sandboxID=7a37710cfebf59892a4d6b3c2e12765b60414f6a7cfb8933a050df609c94e690 Jan 05 09:42:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:16.069171076Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=f0e33b23-1960-44d0-9ec1-fae6f21027c8 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:16.069471871Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=f0e33b23-1960-44d0-9ec1-fae6f21027c8 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:16.070725401Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=8ee816b8-ecb7-4b6d-91d3-f23841971fb5 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:16.070936733Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=8ee816b8-ecb7-4b6d-91d3-f23841971fb5 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:16.072258701Z" level=info msg="Creating container: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy-metrics" id=34eec248-de90-490c-ba4f-f8ae9c1368cf name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:16.072388226Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:16.100739315Z" level=info msg="Created container 543f2356973a019ec622df9cf319f02abc2eac1228b813bef15ed997a118fe6d: openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh/prometheus-adapter" id=6e2cbb62-3354-4a07-9f3b-64777e2272c5 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:16.101236635Z" level=info msg="Starting container: 543f2356973a019ec622df9cf319f02abc2eac1228b813bef15ed997a118fe6d" id=4165ac4e-84c5-41a6-a51d-950c0313ebae name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:16.203668923Z" level=info msg="Started container" PID=7606 containerID=543f2356973a019ec622df9cf319f02abc2eac1228b813bef15ed997a118fe6d description=openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh/prometheus-adapter id=4165ac4e-84c5-41a6-a51d-950c0313ebae name=/runtime.v1.RuntimeService/StartContainer sandboxID=9d9807a6998df2507a4aa5b9bad79e347bc60d62b254989738c927929d827e6e Jan 05 09:42:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:16.360053710Z" level=info msg="Created container d2de249b9456c83b6cb07b3c7fc5010802a974d250e6d15452f354a4100685f6: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-rbac-proxy-self" id=31a02fdf-191d-427c-8f03-8b79907576f3 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:16.360561200Z" level=info msg="Starting container: d2de249b9456c83b6cb07b3c7fc5010802a974d250e6d15452f354a4100685f6" id=ffc8309a-cb48-4428-9e95-7c5517e81e5c name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:16.384486796Z" level=info msg="Started container" PID=7719 containerID=d2de249b9456c83b6cb07b3c7fc5010802a974d250e6d15452f354a4100685f6 description=openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-rbac-proxy-self id=ffc8309a-cb48-4428-9e95-7c5517e81e5c name=/runtime.v1.RuntimeService/StartContainer sandboxID=1e97b2e64ee6d4f76d829e9619da041f24e571e3de5454c48314d411b8770c02 Jan 05 09:42:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:16.396904301Z" level=info msg="Created container 7eca8a735d65a0f0812794e42a36fd300146f8782a35f3df7830a91dfdd49bef: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy-metrics" id=34eec248-de90-490c-ba4f-f8ae9c1368cf name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:16.397611318Z" level=info msg="Starting container: 7eca8a735d65a0f0812794e42a36fd300146f8782a35f3df7830a91dfdd49bef" id=3e893f95-f100-4f3a-b605-e34026e0a9a6 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:16.430580365Z" level=info msg="Started container" PID=7728 containerID=7eca8a735d65a0f0812794e42a36fd300146f8782a35f3df7830a91dfdd49bef description=openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy-metrics id=3e893f95-f100-4f3a-b605-e34026e0a9a6 name=/runtime.v1.RuntimeService/StartContainer sandboxID=fcdbc115e8986f7ba4e1fc9bfa0728fd149a52733e5f0956b91d0b8262b33ca6 Jan 05 09:42:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:16.526044 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" event=&{ID:a40bf20d-67a4-45e7-b701-ca5a75376c1f Type:ContainerStarted Data:543f2356973a019ec622df9cf319f02abc2eac1228b813bef15ed997a118fe6d} Jan 05 09:42:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:16.526082 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" event=&{ID:a40bf20d-67a4-45e7-b701-ca5a75376c1f Type:ContainerStarted Data:9d9807a6998df2507a4aa5b9bad79e347bc60d62b254989738c927929d827e6e} Jan 05 09:42:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:16.529153 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" event=&{ID:7c63c9be-d1ad-44b6-8b33-09af3ee314af Type:ContainerStarted Data:7eca8a735d65a0f0812794e42a36fd300146f8782a35f3df7830a91dfdd49bef} Jan 05 09:42:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:16.529230 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" event=&{ID:7c63c9be-d1ad-44b6-8b33-09af3ee314af Type:ContainerStarted Data:3cbc4c089943552285b7235ee6268116978902d078a993e4f88285e0c16777ce} Jan 05 09:42:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:16.531370 2588 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:42:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:16.533015 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" event=&{ID:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e Type:ContainerStarted Data:d2de249b9456c83b6cb07b3c7fc5010802a974d250e6d15452f354a4100685f6} Jan 05 09:42:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:16.533044 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" event=&{ID:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e Type:ContainerStarted Data:46e4da5180913795c57fd23e927474161e2562ce8f557a4ef7d187e1da8df07c} Jan 05 09:42:16 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:16.555373 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" event=&{ID:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 Type:ContainerStarted Data:0f22074c4c974c98091736b511fa5de52c01939cfdd9697e4eee429370ad1540} Jan 05 09:42:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:17.294853 2588 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:42:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:18.567954 2588 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" Jan 05 09:42:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:22.242094 2588 kubelet.go:2229] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:42:24 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:24.390104 2588 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-2pfzf" Jan 05 09:42:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:27.793984 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:42:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:27.794559005Z" level=info msg="Running pod sandbox: openshift-monitoring/prometheus-k8s-0/POD" id=f2ada4a5-4306-4490-a0a4-0401ebeb37a8 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:27.794626037Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:31.642047126Z" level=error msg="Failed to cleanup (probably retrying): failed to destroy network for pod sandbox k8s_prometheus-k8s-0_openshift-monitoring_4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879_0(04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0): error removing pod openshift-monitoring_prometheus-k8s-0 from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): netplugin failed with no error message: signal: killed" Jan 05 09:42:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:31.642554857Z" level=info msg="Got pod network &{Name:prometheus-k8s-0 Namespace:openshift-monitoring ID:575fca5742eeb550505988169b65e39889d4e9a4a916a669f68c988db3eb4df1 UID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 NetNS:/var/run/netns/1cd4faa3-1dec-4efe-a53a-6ae0a54cc872 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:42:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:31.642589144Z" level=info msg="Adding pod openshift-monitoring_prometheus-k8s-0 to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:42:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:32.281955 2588 kubelet.go:2229] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/alertmanager-main-0" Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:42:33Z [verbose] Add: openshift-monitoring:prometheus-k8s-0:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/1cd4faa3-1dec-4efe-a53a-6ae0a54cc872"}],"ips":[{"version":"4","interface":0,"address":"10.131.0.14/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: I0105 09:42:33.109251 8397 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"prometheus-k8s-0", UID:"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879", APIVersion:"v1", ResourceVersion:"96803", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.14/23] from openshift-sdn Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.153185112Z" level=info msg="Got pod network &{Name:prometheus-k8s-0 Namespace:openshift-monitoring ID:575fca5742eeb550505988169b65e39889d4e9a4a916a669f68c988db3eb4df1 UID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 NetNS:/var/run/netns/1cd4faa3-1dec-4efe-a53a-6ae0a54cc872 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.153360118Z" level=info msg="Checking pod openshift-monitoring_prometheus-k8s-0 for CNI network multus-cni-network (type=multus)" Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:33.158383 2588 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/prometheus-k8s-0] Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:42:33.158595 2588 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ad4cdcf_3061_4f41_9ab3_5f7f06e2b879.slice/crio-575fca5742eeb550505988169b65e39889d4e9a4a916a669f68c988db3eb4df1.scope WatchSource:0}: Error finding container 575fca5742eeb550505988169b65e39889d4e9a4a916a669f68c988db3eb4df1: Status 404 returned error can't find the container with id 575fca5742eeb550505988169b65e39889d4e9a4a916a669f68c988db3eb4df1 Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.164909036Z" level=info msg="Ran pod sandbox 575fca5742eeb550505988169b65e39889d4e9a4a916a669f68c988db3eb4df1 with infra container: openshift-monitoring/prometheus-k8s-0/POD" id=f2ada4a5-4306-4490-a0a4-0401ebeb37a8 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.166841701Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=fbb2f3c1-bf07-495a-a781-147effb5bfa5 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.167033914Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=fbb2f3c1-bf07-495a-a781-147effb5bfa5 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.167951437Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=7b528ca9-874d-4e0f-857e-3ea582544ba9 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.168124430Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=7b528ca9-874d-4e0f-857e-3ea582544ba9 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.168886775Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-0/init-config-reloader" id=c4351146-5cb6-4ea6-b72a-44ce6f37b4bb name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.169004402Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.286801324Z" level=info msg="Created container db2619f5536b755153c7510155fc858ed5569d94d503f7db584520ce0158dd84: openshift-monitoring/prometheus-k8s-0/init-config-reloader" id=c4351146-5cb6-4ea6-b72a-44ce6f37b4bb name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.287296107Z" level=info msg="Starting container: db2619f5536b755153c7510155fc858ed5569d94d503f7db584520ce0158dd84" id=dd558961-bc0f-4abb-a8dc-0afd798954bc name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.296787786Z" level=info msg="Started container" PID=8539 containerID=db2619f5536b755153c7510155fc858ed5569d94d503f7db584520ce0158dd84 description=openshift-monitoring/prometheus-k8s-0/init-config-reloader id=dd558961-bc0f-4abb-a8dc-0afd798954bc name=/runtime.v1.RuntimeService/StartContainer sandboxID=575fca5742eeb550505988169b65e39889d4e9a4a916a669f68c988db3eb4df1 Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:33.599618 2588 generic.go:296] "Generic (PLEG): container finished" podID=4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 containerID="db2619f5536b755153c7510155fc858ed5569d94d503f7db584520ce0158dd84" exitCode=0 Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:33.599654 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event=&{ID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Type:ContainerDied Data:db2619f5536b755153c7510155fc858ed5569d94d503f7db584520ce0158dd84} Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:33.599678 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event=&{ID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Type:ContainerStarted Data:575fca5742eeb550505988169b65e39889d4e9a4a916a669f68c988db3eb4df1} Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.600561576Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:91e7d68b28aa00e209065604b2b9ac1400671d9102581e8b49b37726acd7b6e4" id=c91f7019-e7d1-418b-9d25-89475b91583c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.600767316Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b68b0fa7ea4653b56a91a8c213d456cab2f645207af609598f9877a41585eeae,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:91e7d68b28aa00e209065604b2b9ac1400671d9102581e8b49b37726acd7b6e4],Size_:493220384,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=c91f7019-e7d1-418b-9d25-89475b91583c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.608382357Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:91e7d68b28aa00e209065604b2b9ac1400671d9102581e8b49b37726acd7b6e4" id=fbab119b-0f7d-46b2-9301-78e64c2a6eff name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.608600026Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b68b0fa7ea4653b56a91a8c213d456cab2f645207af609598f9877a41585eeae,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:91e7d68b28aa00e209065604b2b9ac1400671d9102581e8b49b37726acd7b6e4],Size_:493220384,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=fbab119b-0f7d-46b2-9301-78e64c2a6eff name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.611238993Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-0/prometheus" id=7dac334e-8440-49aa-903f-9e02dd4451eb name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.611385501Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.746812311Z" level=info msg="Created container 4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73: openshift-monitoring/prometheus-k8s-0/prometheus" id=7dac334e-8440-49aa-903f-9e02dd4451eb name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.747488740Z" level=info msg="Starting container: 4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73" id=ae837f71-f9a3-48ff-9a88-271f3e8161bd name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.759804757Z" level=info msg="Started container" PID=8596 containerID=4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73 description=openshift-monitoring/prometheus-k8s-0/prometheus id=ae837f71-f9a3-48ff-9a88-271f3e8161bd name=/runtime.v1.RuntimeService/StartContainer sandboxID=575fca5742eeb550505988169b65e39889d4e9a4a916a669f68c988db3eb4df1 Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.773359344Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=5b005a14-dc65-47f4-82e6-c35cf6ceedce name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.773557816Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=5b005a14-dc65-47f4-82e6-c35cf6ceedce name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.774281685Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=1428f6b6-469c-4cf3-9350-5dd53d4d33c9 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.774455022Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=1428f6b6-469c-4cf3-9350-5dd53d4d33c9 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.775281394Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-0/config-reloader" id=33550060-4fef-4db5-a84d-6232782d3245 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.775391922Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.891813903Z" level=info msg="Created container bef5095b779c6d727ff2f590bfd7c28dbad80564fc113cb2c36ebaa3ebaeaedf: openshift-monitoring/prometheus-k8s-0/config-reloader" id=33550060-4fef-4db5-a84d-6232782d3245 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.892435119Z" level=info msg="Starting container: bef5095b779c6d727ff2f590bfd7c28dbad80564fc113cb2c36ebaa3ebaeaedf" id=b4944d60-0256-47e2-9943-cbf6fb5fb7f1 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.901318224Z" level=info msg="Started container" PID=8629 containerID=bef5095b779c6d727ff2f590bfd7c28dbad80564fc113cb2c36ebaa3ebaeaedf description=openshift-monitoring/prometheus-k8s-0/config-reloader id=b4944d60-0256-47e2-9943-cbf6fb5fb7f1 name=/runtime.v1.RuntimeService/StartContainer sandboxID=575fca5742eeb550505988169b65e39889d4e9a4a916a669f68c988db3eb4df1 Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.924247923Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc" id=67a2496d-b247-4d2d-acfa-252c777492ae name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.924490040Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:08ff7a43d9185529520a29e114f0c28b070000b464a44b347240015f3b466488,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc],Size_:368278089,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=67a2496d-b247-4d2d-acfa-252c777492ae name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.925479665Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc" id=56057ed3-9a0d-478b-9224-ab0c72f3dca6 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.925677439Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:08ff7a43d9185529520a29e114f0c28b070000b464a44b347240015f3b466488,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc],Size_:368278089,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=56057ed3-9a0d-478b-9224-ab0c72f3dca6 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.926410175Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-0/thanos-sidecar" id=73faa016-597c-4d86-944e-ed5ca2e9b862 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:33.926532776Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:34.199683398Z" level=info msg="Created container 7a2f8f0ad0be821f8959bfd50621160131d618f2b80fdd07c0192537a6788d3a: openshift-monitoring/prometheus-k8s-0/thanos-sidecar" id=73faa016-597c-4d86-944e-ed5ca2e9b862 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:34.206878312Z" level=info msg="Starting container: 7a2f8f0ad0be821f8959bfd50621160131d618f2b80fdd07c0192537a6788d3a" id=49dc9ea2-3ef6-42ad-a03d-d470e6a71cf1 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:34.280993441Z" level=info msg="Started container" PID=8668 containerID=7a2f8f0ad0be821f8959bfd50621160131d618f2b80fdd07c0192537a6788d3a description=openshift-monitoring/prometheus-k8s-0/thanos-sidecar id=49dc9ea2-3ef6-42ad-a03d-d470e6a71cf1 name=/runtime.v1.RuntimeService/StartContainer sandboxID=575fca5742eeb550505988169b65e39889d4e9a4a916a669f68c988db3eb4df1 Jan 05 09:42:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:34.310048955Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=5d9bdde0-cb5a-43da-a049-9de451ceffa7 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:34.310324938Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=5d9bdde0-cb5a-43da-a049-9de451ceffa7 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:34.311293789Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=f147efc0-5964-4951-bab5-39a751437eeb name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:34.311492850Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=f147efc0-5964-4951-bab5-39a751437eeb name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:34.312593856Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-0/prometheus-proxy" id=f0946338-8f10-45e8-bf95-93f68994dd9d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:34.312719076Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:34.609809 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event=&{ID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Type:ContainerStarted Data:7a2f8f0ad0be821f8959bfd50621160131d618f2b80fdd07c0192537a6788d3a} Jan 05 09:42:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:34.609851 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event=&{ID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Type:ContainerStarted Data:bef5095b779c6d727ff2f590bfd7c28dbad80564fc113cb2c36ebaa3ebaeaedf} Jan 05 09:42:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:34.609867 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event=&{ID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Type:ContainerStarted Data:4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73} Jan 05 09:42:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:34.678386346Z" level=info msg="Created container a988dfd069a4d8b4e201e470b1b75848391f749fd271851262309d39a93dc410: openshift-monitoring/prometheus-k8s-0/prometheus-proxy" id=f0946338-8f10-45e8-bf95-93f68994dd9d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:34.679233254Z" level=info msg="Starting container: a988dfd069a4d8b4e201e470b1b75848391f749fd271851262309d39a93dc410" id=e4a559fa-8cf0-4a20-8da5-bfc2493db46a name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:34.700547981Z" level=info msg="Started container" PID=8744 containerID=a988dfd069a4d8b4e201e470b1b75848391f749fd271851262309d39a93dc410 description=openshift-monitoring/prometheus-k8s-0/prometheus-proxy id=e4a559fa-8cf0-4a20-8da5-bfc2493db46a name=/runtime.v1.RuntimeService/StartContainer sandboxID=575fca5742eeb550505988169b65e39889d4e9a4a916a669f68c988db3eb4df1 Jan 05 09:42:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:34.748992458Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=92bdde35-5281-4505-b664-8439e6917362 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:34.749373639Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=92bdde35-5281-4505-b664-8439e6917362 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:34.750633512Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=b6fc1898-237d-4003-861c-b223c864664c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:34.750820602Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b6fc1898-237d-4003-861c-b223c864664c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:34.752104275Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-0/kube-rbac-proxy" id=7c78c2d7-4319-4e32-8447-9a0036936423 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:34.752254348Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:35.266139119Z" level=info msg="Created container 7de6261b7ebb7d69cfdcbb24569dafb351175e25ae8aecd812d15cab2e644276: openshift-monitoring/prometheus-k8s-0/kube-rbac-proxy" id=7c78c2d7-4319-4e32-8447-9a0036936423 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:35.266828748Z" level=info msg="Starting container: 7de6261b7ebb7d69cfdcbb24569dafb351175e25ae8aecd812d15cab2e644276" id=be3b3284-6784-48b9-aaf9-97ecbeebf319 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:35.301986162Z" level=info msg="Started container" PID=8818 containerID=7de6261b7ebb7d69cfdcbb24569dafb351175e25ae8aecd812d15cab2e644276 description=openshift-monitoring/prometheus-k8s-0/kube-rbac-proxy id=be3b3284-6784-48b9-aaf9-97ecbeebf319 name=/runtime.v1.RuntimeService/StartContainer sandboxID=575fca5742eeb550505988169b65e39889d4e9a4a916a669f68c988db3eb4df1 Jan 05 09:42:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:35.325648041Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=52e58d7d-09a7-43c7-a1e6-5e24b2696803 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:35.325874429Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=52e58d7d-09a7-43c7-a1e6-5e24b2696803 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:35.326661867Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=3a2aed88-a00d-4694-80d1-334d3ec2594e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:35.326844460Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=3a2aed88-a00d-4694-80d1-334d3ec2594e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:35.327886698Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-0/kube-rbac-proxy-thanos" id=f585fd28-9d8e-47df-ae77-b0639f2a4c73 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:35.328000805Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:42:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:35.439695171Z" level=info msg="cleanup sandbox network" Jan 05 09:42:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:35.440088878Z" level=info msg="Got pod network &{Name:prometheus-k8s-0 Namespace:openshift-monitoring ID:04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0 UID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:42:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:35.440346563Z" level=info msg="Deleting pod openshift-monitoring_prometheus-k8s-0 from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:42:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:35.625447 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event=&{ID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Type:ContainerStarted Data:7de6261b7ebb7d69cfdcbb24569dafb351175e25ae8aecd812d15cab2e644276} Jan 05 09:42:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:35.625490 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event=&{ID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Type:ContainerStarted Data:a988dfd069a4d8b4e201e470b1b75848391f749fd271851262309d39a93dc410} Jan 05 09:42:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:35.691388729Z" level=info msg="Created container edb583e71afefdc5a9b0a5e0042237321ce66a47c359a57e1eec8611ac5ed37d: openshift-monitoring/prometheus-k8s-0/kube-rbac-proxy-thanos" id=f585fd28-9d8e-47df-ae77-b0639f2a4c73 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:42:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:35.696821471Z" level=info msg="Starting container: edb583e71afefdc5a9b0a5e0042237321ce66a47c359a57e1eec8611ac5ed37d" id=f951ba0c-d6e5-4be5-a171-c062de15faab name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:42:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:35.750141180Z" level=info msg="Started container" PID=8869 containerID=edb583e71afefdc5a9b0a5e0042237321ce66a47c359a57e1eec8611ac5ed37d description=openshift-monitoring/prometheus-k8s-0/kube-rbac-proxy-thanos id=f951ba0c-d6e5-4be5-a171-c062de15faab name=/runtime.v1.RuntimeService/StartContainer sandboxID=575fca5742eeb550505988169b65e39889d4e9a4a916a669f68c988db3eb4df1 Jan 05 09:42:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:36.496017407Z" level=warning msg="Found defunct process with PID 8106 (haproxy)" Jan 05 09:42:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:36.496134788Z" level=warning msg="Found defunct process with PID 8323 (haproxy)" Jan 05 09:42:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:36.633580 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event=&{ID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Type:ContainerStarted Data:edb583e71afefdc5a9b0a5e0042237321ce66a47c359a57e1eec8611ac5ed37d} Jan 05 09:42:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:37.347513 2588 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:42:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:40.454553 2588 kubelet.go:1343] "Image garbage collection succeeded" Jan 05 09:42:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:40.661357482Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=1cc245a0-95ba-4052-8629-0408bfae73d0 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:42:40.661593804Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=1cc245a0-95ba-4052-8629-0408bfae73d0 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:42:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:42.346481 2588 kubelet.go:2229] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:42:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:42.398542 2588 kubelet.go:2229] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:42:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:42.704053 2588 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-k8s-0" Jan 05 09:42:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:47.307941 2588 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" Jan 05 09:42:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:42:52.633578 2588 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 09:43:20 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:43:20.462292733Z" level=error msg="Failed to cleanup (probably retrying): failed to destroy network for pod sandbox k8s_prometheus-k8s-0_openshift-monitoring_4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879_0(04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0): error removing pod openshift-monitoring_prometheus-k8s-0 from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-monitoring/prometheus-k8s-0/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" Jan 05 09:43:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:43:26.157926741Z" level=info msg="cleanup sandbox network" Jan 05 09:43:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:43:26.158359229Z" level=info msg="Got pod network &{Name:prometheus-k8s-0 Namespace:openshift-monitoring ID:04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0 UID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:43:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:43:26.158537083Z" level=info msg="Deleting pod openshift-monitoring_prometheus-k8s-0 from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.267449 2588 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-kube-storage-version-migrator/migrator-6cd87bdff8-wj8dg] Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.267504 2588 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.271898 2588 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-image-registry/image-registry-64655f6865-jdhwb] Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.271937 2588 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.432774 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-registry-tls\") pod \"image-registry-64655f6865-jdhwb\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " pod="openshift-image-registry/image-registry-64655f6865-jdhwb" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.432817 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-bound-sa-token\") pod \"image-registry-64655f6865-jdhwb\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " pod="openshift-image-registry/image-registry-64655f6865-jdhwb" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.432849 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k224n\" (UniqueName: \"kubernetes.io/projected/08c04948-e189-4227-bece-151bf3576721-kube-api-access-k224n\") pod \"migrator-6cd87bdff8-wj8dg\" (UID: \"08c04948-e189-4227-bece-151bf3576721\") " pod="openshift-kube-storage-version-migrator/migrator-6cd87bdff8-wj8dg" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.432878 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-registry-certificates\") pod \"image-registry-64655f6865-jdhwb\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " pod="openshift-image-registry/image-registry-64655f6865-jdhwb" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.432931 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-image-registry-private-configuration\") pod \"image-registry-64655f6865-jdhwb\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " pod="openshift-image-registry/image-registry-64655f6865-jdhwb" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.432961 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-trusted-ca\") pod \"image-registry-64655f6865-jdhwb\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " pod="openshift-image-registry/image-registry-64655f6865-jdhwb" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.433000 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-installation-pull-secrets\") pod \"image-registry-64655f6865-jdhwb\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " pod="openshift-image-registry/image-registry-64655f6865-jdhwb" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.433037 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-ca-trust-extracted\") pod \"image-registry-64655f6865-jdhwb\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " pod="openshift-image-registry/image-registry-64655f6865-jdhwb" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.433077 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-484qm\" (UniqueName: \"kubernetes.io/projected/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-kube-api-access-484qm\") pod \"image-registry-64655f6865-jdhwb\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " pod="openshift-image-registry/image-registry-64655f6865-jdhwb" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.536920 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-image-registry-private-configuration\") pod \"image-registry-64655f6865-jdhwb\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " pod="openshift-image-registry/image-registry-64655f6865-jdhwb" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.536977 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-trusted-ca\") pod \"image-registry-64655f6865-jdhwb\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " pod="openshift-image-registry/image-registry-64655f6865-jdhwb" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.537020 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-installation-pull-secrets\") pod \"image-registry-64655f6865-jdhwb\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " pod="openshift-image-registry/image-registry-64655f6865-jdhwb" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.537060 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-ca-trust-extracted\") pod \"image-registry-64655f6865-jdhwb\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " pod="openshift-image-registry/image-registry-64655f6865-jdhwb" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.537102 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-484qm\" (UniqueName: \"kubernetes.io/projected/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-kube-api-access-484qm\") pod \"image-registry-64655f6865-jdhwb\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " pod="openshift-image-registry/image-registry-64655f6865-jdhwb" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.537174 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-registry-tls\") pod \"image-registry-64655f6865-jdhwb\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " pod="openshift-image-registry/image-registry-64655f6865-jdhwb" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.537209 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-bound-sa-token\") pod \"image-registry-64655f6865-jdhwb\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " pod="openshift-image-registry/image-registry-64655f6865-jdhwb" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.537246 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-k224n\" (UniqueName: \"kubernetes.io/projected/08c04948-e189-4227-bece-151bf3576721-kube-api-access-k224n\") pod \"migrator-6cd87bdff8-wj8dg\" (UID: \"08c04948-e189-4227-bece-151bf3576721\") " pod="openshift-kube-storage-version-migrator/migrator-6cd87bdff8-wj8dg" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.537297 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-registry-certificates\") pod \"image-registry-64655f6865-jdhwb\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " pod="openshift-image-registry/image-registry-64655f6865-jdhwb" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.538232 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-registry-certificates\") pod \"image-registry-64655f6865-jdhwb\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " pod="openshift-image-registry/image-registry-64655f6865-jdhwb" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.540978 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-image-registry-private-configuration\") pod \"image-registry-64655f6865-jdhwb\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " pod="openshift-image-registry/image-registry-64655f6865-jdhwb" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.541784 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-trusted-ca\") pod \"image-registry-64655f6865-jdhwb\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " pod="openshift-image-registry/image-registry-64655f6865-jdhwb" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.544206 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-installation-pull-secrets\") pod \"image-registry-64655f6865-jdhwb\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " pod="openshift-image-registry/image-registry-64655f6865-jdhwb" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.544509 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-ca-trust-extracted\") pod \"image-registry-64655f6865-jdhwb\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " pod="openshift-image-registry/image-registry-64655f6865-jdhwb" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.547205 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-registry-tls\") pod \"image-registry-64655f6865-jdhwb\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " pod="openshift-image-registry/image-registry-64655f6865-jdhwb" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.587126 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-484qm\" (UniqueName: \"kubernetes.io/projected/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-kube-api-access-484qm\") pod \"image-registry-64655f6865-jdhwb\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " pod="openshift-image-registry/image-registry-64655f6865-jdhwb" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.598231 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-bound-sa-token\") pod \"image-registry-64655f6865-jdhwb\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " pod="openshift-image-registry/image-registry-64655f6865-jdhwb" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.604323 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-k224n\" (UniqueName: \"kubernetes.io/projected/08c04948-e189-4227-bece-151bf3576721-kube-api-access-k224n\") pod \"migrator-6cd87bdff8-wj8dg\" (UID: \"08c04948-e189-4227-bece-151bf3576721\") " pod="openshift-kube-storage-version-migrator/migrator-6cd87bdff8-wj8dg" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.612080 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-6cd87bdff8-wj8dg" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:02.612583417Z" level=info msg="Running pod sandbox: openshift-kube-storage-version-migrator/migrator-6cd87bdff8-wj8dg/POD" id=b1e587b2-9760-46c5-9b6a-5833608b2403 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:02.612640422Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:02.624840 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-64655f6865-jdhwb" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:02.626510810Z" level=info msg="Running pod sandbox: openshift-image-registry/image-registry-64655f6865-jdhwb/POD" id=c25e5943-1be1-47fe-985c-41f1344cb34d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:02.626546835Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:02.630321086Z" level=info msg="Got pod network &{Name:migrator-6cd87bdff8-wj8dg Namespace:openshift-kube-storage-version-migrator ID:943df75e8c120cec9d37e4a8dd05aed3e8f310e67c6a2c33da550e6cdbd653f4 UID:08c04948-e189-4227-bece-151bf3576721 NetNS:/var/run/netns/88f50c67-7c0b-495f-b541-b4d9de0f7858 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:02.630348516Z" level=info msg="Adding pod openshift-kube-storage-version-migrator_migrator-6cd87bdff8-wj8dg to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:02.662232907Z" level=info msg="Got pod network &{Name:image-registry-64655f6865-jdhwb Namespace:openshift-image-registry ID:9640eb4f9700129ebcb44b35913e1a9dfbf8801ccc331fb964566aaba63d21d3 UID:ba5b2420-eec2-4a15-96cc-fb8dd15e86e8 NetNS:/var/run/netns/ddd704d5-9de0-48ff-80c5-4a9c024fe20c Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:44:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:02.662280718Z" level=info msg="Adding pod openshift-image-registry_image-registry-64655f6865-jdhwb to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:04.159737 2588 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-image-registry/image-registry-64655f6865-jdhwb] Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:44:04Z [verbose] Add: openshift-image-registry:image-registry-64655f6865-jdhwb:ba5b2420-eec2-4a15-96cc-fb8dd15e86e8:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/ddd704d5-9de0-48ff-80c5-4a9c024fe20c"}],"ips":[{"version":"4","interface":0,"address":"10.131.0.16/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: I0105 09:44:04.118191 11299 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-image-registry", Name:"image-registry-64655f6865-jdhwb", UID:"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8", APIVersion:"v1", ResourceVersion:"101831", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.16/23] from openshift-sdn Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:04.165037723Z" level=info msg="Got pod network &{Name:image-registry-64655f6865-jdhwb Namespace:openshift-image-registry ID:9640eb4f9700129ebcb44b35913e1a9dfbf8801ccc331fb964566aaba63d21d3 UID:ba5b2420-eec2-4a15-96cc-fb8dd15e86e8 NetNS:/var/run/netns/ddd704d5-9de0-48ff-80c5-4a9c024fe20c Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:04.165207642Z" level=info msg="Checking pod openshift-image-registry_image-registry-64655f6865-jdhwb for CNI network multus-cni-network (type=multus)" Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:44:04.167780 2588 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podba5b2420_eec2_4a15_96cc_fb8dd15e86e8.slice/crio-9640eb4f9700129ebcb44b35913e1a9dfbf8801ccc331fb964566aaba63d21d3.scope WatchSource:0}: Error finding container 9640eb4f9700129ebcb44b35913e1a9dfbf8801ccc331fb964566aaba63d21d3: Status 404 returned error can't find the container with id 9640eb4f9700129ebcb44b35913e1a9dfbf8801ccc331fb964566aaba63d21d3 Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:04.177413213Z" level=info msg="Ran pod sandbox 9640eb4f9700129ebcb44b35913e1a9dfbf8801ccc331fb964566aaba63d21d3 with infra container: openshift-image-registry/image-registry-64655f6865-jdhwb/POD" id=c25e5943-1be1-47fe-985c-41f1344cb34d name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:04.178631203Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8" id=ef705be6-05f8-4df6-87c1-97f2eb8d58b3 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:04.178801856Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:c9e2954d8b9d9e1594d0a16b90086a9297ffd65773a6aa4a64f325e871ce83d1,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8],Size_:363009057,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=ef705be6-05f8-4df6-87c1-97f2eb8d58b3 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:04.179530825Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8" id=6c996ace-bbfb-4029-a9d8-f3ce58715c2c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:04.179739839Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:c9e2954d8b9d9e1594d0a16b90086a9297ffd65773a6aa4a64f325e871ce83d1,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8],Size_:363009057,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=6c996ace-bbfb-4029-a9d8-f3ce58715c2c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:04.180698802Z" level=info msg="Creating container: openshift-image-registry/image-registry-64655f6865-jdhwb/registry" id=f5bc95a5-8ce6-41c6-9db8-21f777916580 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:04.180805135Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:04.336648799Z" level=info msg="Created container 85ffd0a49a25ead6a76fd736bb42c420582a8fb24175f37742eccf7af61d0088: openshift-image-registry/image-registry-64655f6865-jdhwb/registry" id=f5bc95a5-8ce6-41c6-9db8-21f777916580 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:04.337215904Z" level=info msg="Starting container: 85ffd0a49a25ead6a76fd736bb42c420582a8fb24175f37742eccf7af61d0088" id=c99ce19f-a10b-4d16-9f49-c5bb1d400cd8 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:04.346240023Z" level=info msg="Started container" PID=11433 containerID=85ffd0a49a25ead6a76fd736bb42c420582a8fb24175f37742eccf7af61d0088 description=openshift-image-registry/image-registry-64655f6865-jdhwb/registry id=c99ce19f-a10b-4d16-9f49-c5bb1d400cd8 name=/runtime.v1.RuntimeService/StartContainer sandboxID=9640eb4f9700129ebcb44b35913e1a9dfbf8801ccc331fb964566aaba63d21d3 Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:04.726591 2588 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-kube-storage-version-migrator/migrator-6cd87bdff8-wj8dg] Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:44:04Z [verbose] Add: openshift-kube-storage-version-migrator:migrator-6cd87bdff8-wj8dg:08c04948-e189-4227-bece-151bf3576721:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/88f50c67-7c0b-495f-b541-b4d9de0f7858"}],"ips":[{"version":"4","interface":0,"address":"10.131.0.15/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: I0105 09:44:04.699919 11293 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-kube-storage-version-migrator", Name:"migrator-6cd87bdff8-wj8dg", UID:"08c04948-e189-4227-bece-151bf3576721", APIVersion:"v1", ResourceVersion:"101826", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.15/23] from openshift-sdn Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:04.730873323Z" level=info msg="Got pod network &{Name:migrator-6cd87bdff8-wj8dg Namespace:openshift-kube-storage-version-migrator ID:943df75e8c120cec9d37e4a8dd05aed3e8f310e67c6a2c33da550e6cdbd653f4 UID:08c04948-e189-4227-bece-151bf3576721 NetNS:/var/run/netns/88f50c67-7c0b-495f-b541-b4d9de0f7858 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:04.730987013Z" level=info msg="Checking pod openshift-kube-storage-version-migrator_migrator-6cd87bdff8-wj8dg for CNI network multus-cni-network (type=multus)" Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:04.742854526Z" level=info msg="Ran pod sandbox 943df75e8c120cec9d37e4a8dd05aed3e8f310e67c6a2c33da550e6cdbd653f4 with infra container: openshift-kube-storage-version-migrator/migrator-6cd87bdff8-wj8dg/POD" id=b1e587b2-9760-46c5-9b6a-5833608b2403 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:44:04.744608 2588 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08c04948_e189_4227_bece_151bf3576721.slice/crio-943df75e8c120cec9d37e4a8dd05aed3e8f310e67c6a2c33da550e6cdbd653f4.scope WatchSource:0}: Error finding container 943df75e8c120cec9d37e4a8dd05aed3e8f310e67c6a2c33da550e6cdbd653f4: Status 404 returned error can't find the container with id 943df75e8c120cec9d37e4a8dd05aed3e8f310e67c6a2c33da550e6cdbd653f4 Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:04.745415984Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d0336c488437103651d61c3d74b49c13559c7cbc866106918883771b271d0ae9" id=cbfc015d-a5e9-40b6-aa84-0969c8b25562 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:04.745628565Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:65e56f749865ac86fccf0a7cd0a7e5eb6a9964cef9de123547513ceabc7c7051,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d0336c488437103651d61c3d74b49c13559c7cbc866106918883771b271d0ae9],Size_:314684149,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=cbfc015d-a5e9-40b6-aa84-0969c8b25562 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:04.746484332Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d0336c488437103651d61c3d74b49c13559c7cbc866106918883771b271d0ae9" id=9c8d3380-a0c9-48b7-b56b-c99ebaebe84e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:04.746703450Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:65e56f749865ac86fccf0a7cd0a7e5eb6a9964cef9de123547513ceabc7c7051,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d0336c488437103651d61c3d74b49c13559c7cbc866106918883771b271d0ae9],Size_:314684149,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=9c8d3380-a0c9-48b7-b56b-c99ebaebe84e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:04.747514856Z" level=info msg="Creating container: openshift-kube-storage-version-migrator/migrator-6cd87bdff8-wj8dg/migrator" id=06c733ff-b3b4-4e1b-be8b-a35aafc69758 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:04.747630539Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:04.880194 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-6cd87bdff8-wj8dg" event=&{ID:08c04948-e189-4227-bece-151bf3576721 Type:ContainerStarted Data:943df75e8c120cec9d37e4a8dd05aed3e8f310e67c6a2c33da550e6cdbd653f4} Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:04.881353 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-64655f6865-jdhwb" event=&{ID:ba5b2420-eec2-4a15-96cc-fb8dd15e86e8 Type:ContainerStarted Data:85ffd0a49a25ead6a76fd736bb42c420582a8fb24175f37742eccf7af61d0088} Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:04.881384 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-64655f6865-jdhwb" event=&{ID:ba5b2420-eec2-4a15-96cc-fb8dd15e86e8 Type:ContainerStarted Data:9640eb4f9700129ebcb44b35913e1a9dfbf8801ccc331fb964566aaba63d21d3} Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:04.882711 2588 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-64655f6865-jdhwb" Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:04.914912786Z" level=info msg="Created container ae93d6fe80cbb19038b530607e317c9c10827d02079127cc185baa8cd05993b9: openshift-kube-storage-version-migrator/migrator-6cd87bdff8-wj8dg/migrator" id=06c733ff-b3b4-4e1b-be8b-a35aafc69758 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:04.916333720Z" level=info msg="Starting container: ae93d6fe80cbb19038b530607e317c9c10827d02079127cc185baa8cd05993b9" id=61e38f3c-de60-4025-be2a-b7c1d9566962 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:44:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:04.945565846Z" level=info msg="Started container" PID=11495 containerID=ae93d6fe80cbb19038b530607e317c9c10827d02079127cc185baa8cd05993b9 description=openshift-kube-storage-version-migrator/migrator-6cd87bdff8-wj8dg/migrator id=61e38f3c-de60-4025-be2a-b7c1d9566962 name=/runtime.v1.RuntimeService/StartContainer sandboxID=943df75e8c120cec9d37e4a8dd05aed3e8f310e67c6a2c33da550e6cdbd653f4 Jan 05 09:44:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:05.894803 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-6cd87bdff8-wj8dg" event=&{ID:08c04948-e189-4227-bece-151bf3576721 Type:ContainerStarted Data:ae93d6fe80cbb19038b530607e317c9c10827d02079127cc185baa8cd05993b9} Jan 05 09:44:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:11.172436224Z" level=error msg="Failed to cleanup (probably retrying): failed to destroy network for pod sandbox k8s_prometheus-k8s-0_openshift-monitoring_4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879_0(04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0): error removing pod openshift-monitoring_prometheus-k8s-0 from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-monitoring/prometheus-k8s-0/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" Jan 05 09:44:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:19.716425527Z" level=info msg="cleanup sandbox network" Jan 05 09:44:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:19.716816556Z" level=info msg="Got pod network &{Name:prometheus-k8s-0 Namespace:openshift-monitoring ID:04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0 UID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:44:19 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:19.716986466Z" level=info msg="Deleting pod openshift-monitoring_prometheus-k8s-0 from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:44:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:22.630277 2588 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-64655f6865-jdhwb" Jan 05 09:44:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:26.973922 2588 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-multus/cni-sysctl-allowlist-ds-hrwrd] Jan 05 09:44:26 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:26.973977 2588 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:44:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:44:27.001553 2588 watcher.go:93] Error while processing event ("/sys/fs/cgroup/devices/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1166398_2d2c_48bb_8d21_7ca37cefe342.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/devices/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1166398_2d2c_48bb_8d21_7ca37cefe342.slice: no such file or directory Jan 05 09:44:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:27.034904 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcbvl\" (UniqueName: \"kubernetes.io/projected/e1166398-2d2c-48bb-8d21-7ca37cefe342-kube-api-access-lcbvl\") pod \"cni-sysctl-allowlist-ds-hrwrd\" (UID: \"e1166398-2d2c-48bb-8d21-7ca37cefe342\") " pod="openshift-multus/cni-sysctl-allowlist-ds-hrwrd" Jan 05 09:44:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:27.034974 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e1166398-2d2c-48bb-8d21-7ca37cefe342-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-hrwrd\" (UID: \"e1166398-2d2c-48bb-8d21-7ca37cefe342\") " pod="openshift-multus/cni-sysctl-allowlist-ds-hrwrd" Jan 05 09:44:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:27.035034 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/e1166398-2d2c-48bb-8d21-7ca37cefe342-ready\") pod \"cni-sysctl-allowlist-ds-hrwrd\" (UID: \"e1166398-2d2c-48bb-8d21-7ca37cefe342\") " pod="openshift-multus/cni-sysctl-allowlist-ds-hrwrd" Jan 05 09:44:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:27.035071 2588 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e1166398-2d2c-48bb-8d21-7ca37cefe342-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-hrwrd\" (UID: \"e1166398-2d2c-48bb-8d21-7ca37cefe342\") " pod="openshift-multus/cni-sysctl-allowlist-ds-hrwrd" Jan 05 09:44:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:27.136439 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/e1166398-2d2c-48bb-8d21-7ca37cefe342-ready\") pod \"cni-sysctl-allowlist-ds-hrwrd\" (UID: \"e1166398-2d2c-48bb-8d21-7ca37cefe342\") " pod="openshift-multus/cni-sysctl-allowlist-ds-hrwrd" Jan 05 09:44:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:27.136502 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e1166398-2d2c-48bb-8d21-7ca37cefe342-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-hrwrd\" (UID: \"e1166398-2d2c-48bb-8d21-7ca37cefe342\") " pod="openshift-multus/cni-sysctl-allowlist-ds-hrwrd" Jan 05 09:44:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:27.136557 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-lcbvl\" (UniqueName: \"kubernetes.io/projected/e1166398-2d2c-48bb-8d21-7ca37cefe342-kube-api-access-lcbvl\") pod \"cni-sysctl-allowlist-ds-hrwrd\" (UID: \"e1166398-2d2c-48bb-8d21-7ca37cefe342\") " pod="openshift-multus/cni-sysctl-allowlist-ds-hrwrd" Jan 05 09:44:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:27.136599 2588 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e1166398-2d2c-48bb-8d21-7ca37cefe342-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-hrwrd\" (UID: \"e1166398-2d2c-48bb-8d21-7ca37cefe342\") " pod="openshift-multus/cni-sysctl-allowlist-ds-hrwrd" Jan 05 09:44:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:27.137304 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e1166398-2d2c-48bb-8d21-7ca37cefe342-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-hrwrd\" (UID: \"e1166398-2d2c-48bb-8d21-7ca37cefe342\") " pod="openshift-multus/cni-sysctl-allowlist-ds-hrwrd" Jan 05 09:44:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:27.137555 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/e1166398-2d2c-48bb-8d21-7ca37cefe342-ready\") pod \"cni-sysctl-allowlist-ds-hrwrd\" (UID: \"e1166398-2d2c-48bb-8d21-7ca37cefe342\") " pod="openshift-multus/cni-sysctl-allowlist-ds-hrwrd" Jan 05 09:44:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:27.137629 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e1166398-2d2c-48bb-8d21-7ca37cefe342-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-hrwrd\" (UID: \"e1166398-2d2c-48bb-8d21-7ca37cefe342\") " pod="openshift-multus/cni-sysctl-allowlist-ds-hrwrd" Jan 05 09:44:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:27.178565 2588 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcbvl\" (UniqueName: \"kubernetes.io/projected/e1166398-2d2c-48bb-8d21-7ca37cefe342-kube-api-access-lcbvl\") pod \"cni-sysctl-allowlist-ds-hrwrd\" (UID: \"e1166398-2d2c-48bb-8d21-7ca37cefe342\") " pod="openshift-multus/cni-sysctl-allowlist-ds-hrwrd" Jan 05 09:44:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:27.302517 2588 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-hrwrd" Jan 05 09:44:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:27.303336231Z" level=info msg="Running pod sandbox: openshift-multus/cni-sysctl-allowlist-ds-hrwrd/POD" id=b6df0c75-9f5f-40dd-a561-ffdcfc66e6cb name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:44:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:27.303406391Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:44:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:27.341124480Z" level=info msg="Got pod network &{Name:cni-sysctl-allowlist-ds-hrwrd Namespace:openshift-multus ID:b7acbcee674c440f0bf1b02b910456c58b573f0e935553e9c34a3510945d22ae UID:e1166398-2d2c-48bb-8d21-7ca37cefe342 NetNS:/var/run/netns/6ad09ef5-2f23-47e3-afe2-734fa0458f87 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:44:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:27.341187887Z" level=info msg="Adding pod openshift-multus_cni-sysctl-allowlist-ds-hrwrd to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:44:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:27.637396177Z" level=warning msg="Found defunct process with PID 12102 (haproxy)" Jan 05 09:44:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:31.733172 2588 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-multus/cni-sysctl-allowlist-ds-hrwrd] Jan 05 09:44:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:44:28Z [verbose] Add: openshift-multus:cni-sysctl-allowlist-ds-hrwrd:e1166398-2d2c-48bb-8d21-7ca37cefe342:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/6ad09ef5-2f23-47e3-afe2-734fa0458f87"}],"ips":[{"version":"4","interface":0,"address":"10.131.0.17/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:44:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: I0105 09:44:28.736158 12560 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-multus", Name:"cni-sysctl-allowlist-ds-hrwrd", UID:"e1166398-2d2c-48bb-8d21-7ca37cefe342", APIVersion:"v1", ResourceVersion:"102728", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.17/23] from openshift-sdn Jan 05 09:44:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:31.734749046Z" level=info msg="Got pod network &{Name:cni-sysctl-allowlist-ds-hrwrd Namespace:openshift-multus ID:b7acbcee674c440f0bf1b02b910456c58b573f0e935553e9c34a3510945d22ae UID:e1166398-2d2c-48bb-8d21-7ca37cefe342 NetNS:/var/run/netns/6ad09ef5-2f23-47e3-afe2-734fa0458f87 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:44:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:31.734897271Z" level=info msg="Checking pod openshift-multus_cni-sysctl-allowlist-ds-hrwrd for CNI network multus-cni-network (type=multus)" Jan 05 09:44:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:44:31.742235 2588 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1166398_2d2c_48bb_8d21_7ca37cefe342.slice/crio-b7acbcee674c440f0bf1b02b910456c58b573f0e935553e9c34a3510945d22ae.scope WatchSource:0}: Error finding container b7acbcee674c440f0bf1b02b910456c58b573f0e935553e9c34a3510945d22ae: Status 404 returned error can't find the container with id b7acbcee674c440f0bf1b02b910456c58b573f0e935553e9c34a3510945d22ae Jan 05 09:44:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:31.745911812Z" level=info msg="Ran pod sandbox b7acbcee674c440f0bf1b02b910456c58b573f0e935553e9c34a3510945d22ae with infra container: openshift-multus/cni-sysctl-allowlist-ds-hrwrd/POD" id=b6df0c75-9f5f-40dd-a561-ffdcfc66e6cb name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:44:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:31.748415507Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=25b56d58-df07-4ec2-a8d4-22ce6de101ba name=/runtime.v1.ImageService/ImageStatus Jan 05 09:44:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:31.748621071Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=25b56d58-df07-4ec2-a8d4-22ce6de101ba name=/runtime.v1.ImageService/ImageStatus Jan 05 09:44:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:31.755211364Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=605ac932-3ee3-45e5-8b4e-067757b80b50 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:44:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:31.755421324Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=605ac932-3ee3-45e5-8b4e-067757b80b50 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:44:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:31.766411295Z" level=info msg="Creating container: openshift-multus/cni-sysctl-allowlist-ds-hrwrd/kube-multus-additional-cni-plugins" id=90d1ba63-dc27-41ce-bb69-7bc702a79093 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:44:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:31.766521773Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:44:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:31.945383105Z" level=info msg="Created container 670e670ccf20be64b02c12b1075000c60fafc29d4c8965aee4edaaf717c14ced: openshift-multus/cni-sysctl-allowlist-ds-hrwrd/kube-multus-additional-cni-plugins" id=90d1ba63-dc27-41ce-bb69-7bc702a79093 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:44:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:31.945803662Z" level=info msg="Starting container: 670e670ccf20be64b02c12b1075000c60fafc29d4c8965aee4edaaf717c14ced" id=343456ee-651b-4bb4-8919-a076e28c1b1a name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:44:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:31.953821016Z" level=info msg="Started container" PID=12721 containerID=670e670ccf20be64b02c12b1075000c60fafc29d4c8965aee4edaaf717c14ced description=openshift-multus/cni-sysctl-allowlist-ds-hrwrd/kube-multus-additional-cni-plugins id=343456ee-651b-4bb4-8919-a076e28c1b1a name=/runtime.v1.RuntimeService/StartContainer sandboxID=b7acbcee674c440f0bf1b02b910456c58b573f0e935553e9c34a3510945d22ae Jan 05 09:44:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:31.967511 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-hrwrd" event=&{ID:e1166398-2d2c-48bb-8d21-7ca37cefe342 Type:ContainerStarted Data:670e670ccf20be64b02c12b1075000c60fafc29d4c8965aee4edaaf717c14ced} Jan 05 09:44:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:31.967544 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-hrwrd" event=&{ID:e1166398-2d2c-48bb-8d21-7ca37cefe342 Type:ContainerStarted Data:b7acbcee674c440f0bf1b02b910456c58b573f0e935553e9c34a3510945d22ae} Jan 05 09:44:31 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:31.967637 2588 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-multus/cni-sysctl-allowlist-ds-hrwrd" Jan 05 09:44:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:33.006395 2588 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-multus/cni-sysctl-allowlist-ds-hrwrd" Jan 05 09:44:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:36.206181 2588 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-multus/cni-sysctl-allowlist-ds-hrwrd] Jan 05 09:44:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:36.206392 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-multus/cni-sysctl-allowlist-ds-hrwrd" podUID=e1166398-2d2c-48bb-8d21-7ca37cefe342 containerName="kube-multus-additional-cni-plugins" containerID="cri-o://670e670ccf20be64b02c12b1075000c60fafc29d4c8965aee4edaaf717c14ced" gracePeriod=30 Jan 05 09:44:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:36.207014629Z" level=info msg="Stopping container: 670e670ccf20be64b02c12b1075000c60fafc29d4c8965aee4edaaf717c14ced (timeout: 30s)" id=f56b897a-0771-42f9-9cbb-227a9aee7a07 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:36.491634980Z" level=warning msg="Found defunct process with PID 12304 (haproxy)" Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.376688 2588 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256] Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.376909 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" podUID=9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 containerName="kube-rbac-proxy-main" containerID="cri-o://10f645c424bd560dd410724c8e884a10049cfe720e59016dcb3f8d51623b3ae6" gracePeriod=30 Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.377168 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" podUID=9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 containerName="openshift-state-metrics" containerID="cri-o://0f22074c4c974c98091736b511fa5de52c01939cfdd9697e4eee429370ad1540" gracePeriod=30 Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.377237 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" podUID=9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 containerName="kube-rbac-proxy-self" containerID="cri-o://489142621cb0989124915d3c9efa7b65c6ff3af9ed775b03b9f887e6ad7d1d8b" gracePeriod=30 Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:46.378224301Z" level=info msg="Stopping container: 489142621cb0989124915d3c9efa7b65c6ff3af9ed775b03b9f887e6ad7d1d8b (timeout: 30s)" id=b51ecc2a-0b68-4a4c-8868-72865a398b62 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:46.378500200Z" level=info msg="Stopping container: 10f645c424bd560dd410724c8e884a10049cfe720e59016dcb3f8d51623b3ae6 (timeout: 30s)" id=833c265e-f5eb-4f1d-89ec-66803605a5ff name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:46.378671482Z" level=info msg="Stopping container: 0f22074c4c974c98091736b511fa5de52c01939cfdd9697e4eee429370ad1540 (timeout: 30s)" id=519ff9a4-8e19-4a48-acb5-fa17b549a9e7 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.395301 2588 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-monitoring/kube-state-metrics-6798b548b9-xl88q] Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.395533 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" podUID=2ecd2e2c-ab52-42e0-8fde-64e45804cc9e containerName="kube-state-metrics" containerID="cri-o://987ad82f0f8b3760e232b876a53eb03f7fa455eadddd9e240e5e8be4f57bf369" gracePeriod=30 Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.395861 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" podUID=2ecd2e2c-ab52-42e0-8fde-64e45804cc9e containerName="kube-rbac-proxy-self" containerID="cri-o://d2de249b9456c83b6cb07b3c7fc5010802a974d250e6d15452f354a4100685f6" gracePeriod=30 Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.395970 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" podUID=2ecd2e2c-ab52-42e0-8fde-64e45804cc9e containerName="kube-rbac-proxy-main" containerID="cri-o://46e4da5180913795c57fd23e927474161e2562ce8f557a4ef7d187e1da8df07c" gracePeriod=30 Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:46.397659305Z" level=info msg="Stopping container: 46e4da5180913795c57fd23e927474161e2562ce8f557a4ef7d187e1da8df07c (timeout: 30s)" id=3711d209-e841-491f-93ab-0c6d4dc36307 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:46.397888360Z" level=info msg="Stopping container: 987ad82f0f8b3760e232b876a53eb03f7fa455eadddd9e240e5e8be4f57bf369 (timeout: 30s)" id=6a62ee9d-a00c-498d-9668-56d1b3d45a17 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:46.398039759Z" level=info msg="Stopping container: d2de249b9456c83b6cb07b3c7fc5010802a974d250e6d15452f354a4100685f6 (timeout: 30s)" id=8676d75b-e340-4481-b963-f21b1de22117 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.401691 2588 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l] Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.401727 2588 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-ingress/router-default-5cf6bbdff9-tnffs] Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.401831 2588 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-kube-storage-version-migrator/migrator-6cd87bdff8-wj8dg] Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.401874 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" podUID=c3b10ff4-a243-4a4e-b27d-f613e258666d containerName="router" containerID="cri-o://8eaf1cd2c43d71002a86869a186ab94e90ccc29be62720e01c85c11de8ae9285" gracePeriod=3600 Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.401942 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-kube-storage-version-migrator/migrator-6cd87bdff8-wj8dg" podUID=08c04948-e189-4227-bece-151bf3576721 containerName="migrator" containerID="cri-o://ae93d6fe80cbb19038b530607e317c9c10827d02079127cc185baa8cd05993b9" gracePeriod=30 Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.402126 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" podUID=7c63c9be-d1ad-44b6-8b33-09af3ee314af containerName="thanos-query" containerID="cri-o://578cce0ed1aa4795c3b6d8648ecf4b937233d871e6a027eb233bec20d4333c1f" gracePeriod=120 Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.402241 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" podUID=7c63c9be-d1ad-44b6-8b33-09af3ee314af containerName="kube-rbac-proxy-metrics" containerID="cri-o://7eca8a735d65a0f0812794e42a36fd300146f8782a35f3df7830a91dfdd49bef" gracePeriod=120 Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.402372 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" podUID=7c63c9be-d1ad-44b6-8b33-09af3ee314af containerName="kube-rbac-proxy-rules" containerID="cri-o://3cbc4c089943552285b7235ee6268116978902d078a993e4f88285e0c16777ce" gracePeriod=120 Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.402432 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" podUID=7c63c9be-d1ad-44b6-8b33-09af3ee314af containerName="prom-label-proxy" containerID="cri-o://b7574d11ea7895beed8a146040a8adb91835493e426424dd8e461a93b6f7677b" gracePeriod=120 Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.402487 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" podUID=7c63c9be-d1ad-44b6-8b33-09af3ee314af containerName="kube-rbac-proxy" containerID="cri-o://c507c8393ab5e7e8bb6dbc97bcabc3f2598889f9b13fbee87614f2fee17b1e9d" gracePeriod=120 Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.402542 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" podUID=7c63c9be-d1ad-44b6-8b33-09af3ee314af containerName="oauth-proxy" containerID="cri-o://76d0bbfe3b04b3fb31c951bf14c59dddccfd473352dd9388eeee0e7e51e96358" gracePeriod=120 Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:46.403406235Z" level=info msg="Stopping container: 76d0bbfe3b04b3fb31c951bf14c59dddccfd473352dd9388eeee0e7e51e96358 (timeout: 120s)" id=73a8add1-7210-497e-82ab-92bec95e4994 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.416204 2588 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-image-registry/image-registry-64655f6865-jdhwb] Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:46.416998242Z" level=info msg="Stopping container: 8eaf1cd2c43d71002a86869a186ab94e90ccc29be62720e01c85c11de8ae9285 (timeout: 3600s)" id=b7fc2463-15a0-46b6-a852-a08443a134ca name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:46.417181340Z" level=info msg="Stopping container: ae93d6fe80cbb19038b530607e317c9c10827d02079127cc185baa8cd05993b9 (timeout: 30s)" id=c98aad84-271a-4804-91cf-0d75e9c22d2b name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:46.417314539Z" level=info msg="Stopping container: 578cce0ed1aa4795c3b6d8648ecf4b937233d871e6a027eb233bec20d4333c1f (timeout: 120s)" id=ff1718a5-8bb4-4a7d-a645-696948506a14 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:46.417428389Z" level=info msg="Stopping container: 7eca8a735d65a0f0812794e42a36fd300146f8782a35f3df7830a91dfdd49bef (timeout: 120s)" id=b9a42aed-536b-4fe5-b0bb-ea6cdcb00b56 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:46.417539398Z" level=info msg="Stopping container: 3cbc4c089943552285b7235ee6268116978902d078a993e4f88285e0c16777ce (timeout: 120s)" id=f7eba895-6408-415b-93f6-80b65c43f505 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:46.417655069Z" level=info msg="Stopping container: b7574d11ea7895beed8a146040a8adb91835493e426424dd8e461a93b6f7677b (timeout: 120s)" id=d72ba8fb-0693-41cd-a8e0-322935bd0425 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:46.417769560Z" level=info msg="Stopping container: c507c8393ab5e7e8bb6dbc97bcabc3f2598889f9b13fbee87614f2fee17b1e9d (timeout: 120s)" id=f0abf1ee-7cb3-4bd5-adca-f95deac91281 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.441245 2588 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh] Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.441454 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" podUID=a40bf20d-67a4-45e7-b701-ca5a75376c1f containerName="prometheus-adapter" containerID="cri-o://543f2356973a019ec622df9cf319f02abc2eac1228b813bef15ed997a118fe6d" gracePeriod=30 Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.443794 2588 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn] Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.443938 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" podUID=88745ab7-efe7-42c0-a2e8-75591518333d containerName="prometheus-operator-admission-webhook" containerID="cri-o://0b999d4ad314ba518c86d2e8df430b23bade34adf813f2325e76c8d1d79e1b52" gracePeriod=30 Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:46.444352129Z" level=info msg="Stopping container: 543f2356973a019ec622df9cf319f02abc2eac1228b813bef15ed997a118fe6d (timeout: 30s)" id=5c8a419b-b6c3-4f78-9e29-6480a77d48f7 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:46.444486247Z" level=info msg="Stopping container: 0b999d4ad314ba518c86d2e8df430b23bade34adf813f2325e76c8d1d79e1b52 (timeout: 30s)" id=7a4ea9f8-6d2f-4f99-9c80-21292490c82f name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.462735 2588 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-monitoring/alertmanager-main-0] Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.462773 2588 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-monitoring/prometheus-k8s-0] Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.463035 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/prometheus-k8s-0" podUID=4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 containerName="prometheus" containerID="cri-o://4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73" gracePeriod=600 Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.463341 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-0" podUID=05c9a6ea-140a-49a2-bde5-f220b05aa252 containerName="alertmanager" containerID="cri-o://29d242ba69edfb0ca301e890f9169a7355a0889dbda9e344277448360467524b" gracePeriod=120 Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.463461 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/prometheus-k8s-0" podUID=4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 containerName="kube-rbac-proxy-thanos" containerID="cri-o://edb583e71afefdc5a9b0a5e0042237321ce66a47c359a57e1eec8611ac5ed37d" gracePeriod=600 Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.463520 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/prometheus-k8s-0" podUID=4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 containerName="kube-rbac-proxy" containerID="cri-o://7de6261b7ebb7d69cfdcbb24569dafb351175e25ae8aecd812d15cab2e644276" gracePeriod=600 Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.463571 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/prometheus-k8s-0" podUID=4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 containerName="prometheus-proxy" containerID="cri-o://a988dfd069a4d8b4e201e470b1b75848391f749fd271851262309d39a93dc410" gracePeriod=600 Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.463624 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/prometheus-k8s-0" podUID=4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 containerName="thanos-sidecar" containerID="cri-o://7a2f8f0ad0be821f8959bfd50621160131d618f2b80fdd07c0192537a6788d3a" gracePeriod=600 Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.463722 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/prometheus-k8s-0" podUID=4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 containerName="config-reloader" containerID="cri-o://bef5095b779c6d727ff2f590bfd7c28dbad80564fc113cb2c36ebaa3ebaeaedf" gracePeriod=600 Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.463846 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-0" podUID=05c9a6ea-140a-49a2-bde5-f220b05aa252 containerName="prom-label-proxy" containerID="cri-o://4a95d41273862399b3cca16471b999defcccef3ceb8402d7cfb798378ffc8cab" gracePeriod=120 Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.463924 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-0" podUID=05c9a6ea-140a-49a2-bde5-f220b05aa252 containerName="kube-rbac-proxy-metric" containerID="cri-o://4c970eddded2809e2e506cfb0b207a19b375bc49b3a1f9bfd58b202b7c1f6ebd" gracePeriod=120 Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.463991 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-0" podUID=05c9a6ea-140a-49a2-bde5-f220b05aa252 containerName="kube-rbac-proxy" containerID="cri-o://53a26d2c619d3ef8230bda385bc9245dc43ada80f94fe54f1f79b82763d2f7a8" gracePeriod=120 Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.464047 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-0" podUID=05c9a6ea-140a-49a2-bde5-f220b05aa252 containerName="alertmanager-proxy" containerID="cri-o://85c843961a0e8a2283e240b3c350b3d3823e13fbe3cb8f291a0e809df84275e7" gracePeriod=120 Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.464181 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/alertmanager-main-0" podUID=05c9a6ea-140a-49a2-bde5-f220b05aa252 containerName="config-reloader" containerID="cri-o://df204cb60308dd066ca32f1f38b7a71d29064341d03539fedabfb840583c8a96" gracePeriod=120 Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:46.475603285Z" level=info msg="Stopping container: df204cb60308dd066ca32f1f38b7a71d29064341d03539fedabfb840583c8a96 (timeout: 120s)" id=8c4de4ee-9f4e-48d6-82db-993826b53ba0 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:46.475779152Z" level=info msg="Stopping container: 4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73 (timeout: 600s)" id=e214be64-8428-42f3-a1a7-d93549e928bd name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:46.475884430Z" level=info msg="Stopping container: 29d242ba69edfb0ca301e890f9169a7355a0889dbda9e344277448360467524b (timeout: 120s)" id=95d74e1c-0308-47ba-84f8-0cd0bea52eb7 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:46.475993346Z" level=info msg="Stopping container: edb583e71afefdc5a9b0a5e0042237321ce66a47c359a57e1eec8611ac5ed37d (timeout: 600s)" id=d0a21dd9-1bdc-4b8c-886c-9c65d6e1083d name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:46.476107068Z" level=info msg="Stopping container: 7de6261b7ebb7d69cfdcbb24569dafb351175e25ae8aecd812d15cab2e644276 (timeout: 600s)" id=4964390b-5cd2-4351-9b78-efd30ead6a0e name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:46.476239370Z" level=info msg="Stopping container: a988dfd069a4d8b4e201e470b1b75848391f749fd271851262309d39a93dc410 (timeout: 600s)" id=1f923910-83f5-4df2-9ce0-35766e81c5c8 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:46.476372988Z" level=info msg="Stopping container: 7a2f8f0ad0be821f8959bfd50621160131d618f2b80fdd07c0192537a6788d3a (timeout: 600s)" id=766fa700-b41c-49cf-aea9-8c2e9370f442 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:46.476489533Z" level=info msg="Stopping container: bef5095b779c6d727ff2f590bfd7c28dbad80564fc113cb2c36ebaa3ebaeaedf (timeout: 600s)" id=6bbd4497-9411-45f4-8746-e17903cae2a1 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:46.476597247Z" level=info msg="Stopping container: 4a95d41273862399b3cca16471b999defcccef3ceb8402d7cfb798378ffc8cab (timeout: 120s)" id=9e418dce-f5e3-4d50-8595-8e568a1c690a name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:46.476716620Z" level=info msg="Stopping container: 4c970eddded2809e2e506cfb0b207a19b375bc49b3a1f9bfd58b202b7c1f6ebd (timeout: 120s)" id=d5e4f5ab-0f11-4cf0-ad26-e8c0fc688e07 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:46.476831847Z" level=info msg="Stopping container: 53a26d2c619d3ef8230bda385bc9245dc43ada80f94fe54f1f79b82763d2f7a8 (timeout: 120s)" id=d874a7ab-fbe5-48dd-add0-42f05d3f2cfd name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:46.476939146Z" level=info msg="Stopping container: 85c843961a0e8a2283e240b3c350b3d3823e13fbe3cb8f291a0e809df84275e7 (timeout: 120s)" id=91d89d57-8d05-4c54-81cd-a02c2316e92f name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.642978 2588 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-monitoring/telemeter-client-5598c645c4-mftdv] Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.643286 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" podUID=fc45e847-8f55-4c48-8c48-4bfee2419dca containerName="telemeter-client" containerID="cri-o://2c1936a19655345a892373f86a623f7409e46829cbb1adeca75e2879d268aa08" gracePeriod=30 Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.643446 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" podUID=fc45e847-8f55-4c48-8c48-4bfee2419dca containerName="kube-rbac-proxy" containerID="cri-o://91947bee1cff242d8ff480a4ba98a9c65dd41326ef2bfc74ccc1a969906e3b00" gracePeriod=30 Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:46.643530 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" podUID=fc45e847-8f55-4c48-8c48-4bfee2419dca containerName="reload" containerID="cri-o://2952a168f0605321dd4ff798bf4481d0ede73526e359849e9e7039226b853fab" gracePeriod=30 Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:46.652834851Z" level=info msg="Stopping container: 2952a168f0605321dd4ff798bf4481d0ede73526e359849e9e7039226b853fab (timeout: 30s)" id=1d534ed0-b6ef-4c25-a69f-14b99b66d67e name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:46.653003699Z" level=info msg="Stopping container: 2c1936a19655345a892373f86a623f7409e46829cbb1adeca75e2879d268aa08 (timeout: 30s)" id=d0544cb6-f40d-42c0-975f-c8fcf50b2494 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:46.653112607Z" level=info msg="Stopping container: 91947bee1cff242d8ff480a4ba98a9c65dd41326ef2bfc74ccc1a969906e3b00 (timeout: 30s)" id=96cbbb48-6558-44d6-8e6c-e906e947dba2 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:47.296845 2588 patch_prober.go:29] interesting pod/prometheus-adapter-5894cc86c7-qhmbh container/prometheus-adapter namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.131.0.13:6443/readyz\": dial tcp 10.131.0.13:6443: connect: connection refused" start-of-body= Jan 05 09:44:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:47.296924 2588 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" podUID=a40bf20d-67a4-45e7-b701-ca5a75376c1f containerName="prometheus-adapter" probeResult=failure output="Get \"https://10.131.0.13:6443/readyz\": dial tcp 10.131.0.13:6443: connect: connection refused" Jan 05 09:44:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:47.313055410Z" level=info msg="Stopped container 76d0bbfe3b04b3fb31c951bf14c59dddccfd473352dd9388eeee0e7e51e96358: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/oauth-proxy" id=73a8add1-7210-497e-82ab-92bec95e4994 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:44:47.352365 2588 remote_runtime.go:734] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73 is running failed: open /proc/8596/stat: no such file or directory: container process not found" containerID="4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73" cmd=[sh -c if [ -x "$(command -v curl)" ]; then exec curl --fail http://localhost:9090/-/ready; elif [ -x "$(command -v wget)" ]; then exec wget -q -O /dev/null http://localhost:9090/-/ready; else exit 1; fi] Jan 05 09:44:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:44:47.352725 2588 remote_runtime.go:734] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73 is running failed: open /proc/8596/stat: no such file or directory: container process not found" containerID="4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73" cmd=[sh -c if [ -x "$(command -v curl)" ]; then exec curl --fail http://localhost:9090/-/ready; elif [ -x "$(command -v wget)" ]; then exec wget -q -O /dev/null http://localhost:9090/-/ready; else exit 1; fi] Jan 05 09:44:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:44:47.352984 2588 remote_runtime.go:734] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73 is running failed: open /proc/8596/stat: no such file or directory: container process not found" containerID="4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73" cmd=[sh -c if [ -x "$(command -v curl)" ]; then exec curl --fail http://localhost:9090/-/ready; elif [ -x "$(command -v wget)" ]; then exec wget -q -O /dev/null http://localhost:9090/-/ready; else exit 1; fi] Jan 05 09:44:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:44:47.353027 2588 prober.go:111] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73 is running failed: open /proc/8596/stat: no such file or directory: container process not found" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID=4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 containerName="prometheus" Jan 05 09:44:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:47.377445 2588 patch_prober.go:29] interesting pod/thanos-querier-6f4f5fb7cf-rdm9l container/oauth-proxy namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.131.0.7:9091/-/ready\": dial tcp 10.131.0.7:9091: connect: connection refused" start-of-body= Jan 05 09:44:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:47.377498 2588 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" podUID=7c63c9be-d1ad-44b6-8b33-09af3ee314af containerName="oauth-proxy" probeResult=failure output="Get \"https://10.131.0.7:9091/-/ready\": dial tcp 10.131.0.7:9091: connect: connection refused" Jan 05 09:44:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:47.396817536Z" level=info msg="Stopped container 578cce0ed1aa4795c3b6d8648ecf4b937233d871e6a027eb233bec20d4333c1f: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/thanos-query" id=ff1718a5-8bb4-4a7d-a645-696948506a14 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:47.475710407Z" level=info msg="Stopped container ae93d6fe80cbb19038b530607e317c9c10827d02079127cc185baa8cd05993b9: openshift-kube-storage-version-migrator/migrator-6cd87bdff8-wj8dg/migrator" id=c98aad84-271a-4804-91cf-0d75e9c22d2b name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:47.476902128Z" level=info msg="Stopping pod sandbox: 943df75e8c120cec9d37e4a8dd05aed3e8f310e67c6a2c33da550e6cdbd653f4" id=8663b4f2-8a78-4721-abda-44407668d7e3 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:44:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:47.486721671Z" level=info msg="Got pod network &{Name:migrator-6cd87bdff8-wj8dg Namespace:openshift-kube-storage-version-migrator ID:943df75e8c120cec9d37e4a8dd05aed3e8f310e67c6a2c33da550e6cdbd653f4 UID:08c04948-e189-4227-bece-151bf3576721 NetNS:/var/run/netns/88f50c67-7c0b-495f-b541-b4d9de0f7858 Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:44:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:47.486961977Z" level=info msg="Deleting pod openshift-kube-storage-version-migrator_migrator-6cd87bdff8-wj8dg from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:44:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:47.789229092Z" level=info msg="Stopped container 987ad82f0f8b3760e232b876a53eb03f7fa455eadddd9e240e5e8be4f57bf369: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-state-metrics" id=6a62ee9d-a00c-498d-9668-56d1b3d45a17 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:47.832939775Z" level=info msg="Stopped container b7574d11ea7895beed8a146040a8adb91835493e426424dd8e461a93b6f7677b: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/prom-label-proxy" id=d72ba8fb-0693-41cd-a8e0-322935bd0425 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:47.856632086Z" level=info msg="Stopped container 0f22074c4c974c98091736b511fa5de52c01939cfdd9697e4eee429370ad1540: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/openshift-state-metrics" id=519ff9a4-8e19-4a48-acb5-fa17b549a9e7 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:47.897993749Z" level=info msg="Stopped container c507c8393ab5e7e8bb6dbc97bcabc3f2598889f9b13fbee87614f2fee17b1e9d: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy" id=f0abf1ee-7cb3-4bd5-adca-f95deac91281 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:47.923427392Z" level=info msg="Stopped container 3cbc4c089943552285b7235ee6268116978902d078a993e4f88285e0c16777ce: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy-rules" id=f7eba895-6408-415b-93f6-80b65c43f505 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:47.928448652Z" level=info msg="Stopped container 0b999d4ad314ba518c86d2e8df430b23bade34adf813f2325e76c8d1d79e1b52: openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn/prometheus-operator-admission-webhook" id=7a4ea9f8-6d2f-4f99-9c80-21292490c82f name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:47.928878903Z" level=info msg="Stopping pod sandbox: e10e02d14faf7624d78743e0d1899ddf64963db130ecf653b0849bcde995a72c" id=85691975-3563-4104-a900-da620801d8b9 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:44:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:47.929342645Z" level=info msg="Got pod network &{Name:prometheus-operator-admission-webhook-598468bb49-94dsn Namespace:openshift-monitoring ID:e10e02d14faf7624d78743e0d1899ddf64963db130ecf653b0849bcde995a72c UID:88745ab7-efe7-42c0-a2e8-75591518333d NetNS:/var/run/netns/67866f67-3eb9-4569-aa7e-69d12f039663 Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:44:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:47.929498597Z" level=info msg="Deleting pod openshift-monitoring_prometheus-operator-admission-webhook-598468bb49-94dsn from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:44:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:44:47Z [verbose] Del: openshift-kube-storage-version-migrator:migrator-6cd87bdff8-wj8dg:08c04948-e189-4227-bece-151bf3576721:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:44:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:47.944936252Z" level=info msg="Stopped container df204cb60308dd066ca32f1f38b7a71d29064341d03539fedabfb840583c8a96: openshift-monitoring/alertmanager-main-0/config-reloader" id=8c4de4ee-9f4e-48d6-82db-993826b53ba0 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:47.963009115Z" level=info msg="Stopped container 4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73: openshift-monitoring/prometheus-k8s-0/prometheus" id=e214be64-8428-42f3-a1a7-d93549e928bd name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:47.986669878Z" level=info msg="Stopped container 4a95d41273862399b3cca16471b999defcccef3ceb8402d7cfb798378ffc8cab: openshift-monitoring/alertmanager-main-0/prom-label-proxy" id=9e418dce-f5e3-4d50-8595-8e568a1c690a name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.007821950Z" level=info msg="Stopped container 2c1936a19655345a892373f86a623f7409e46829cbb1adeca75e2879d268aa08: openshift-monitoring/telemeter-client-5598c645c4-mftdv/telemeter-client" id=d0544cb6-f40d-42c0-975f-c8fcf50b2494 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.053020406Z" level=info msg="Stopped container 53a26d2c619d3ef8230bda385bc9245dc43ada80f94fe54f1f79b82763d2f7a8: openshift-monitoring/alertmanager-main-0/kube-rbac-proxy" id=d874a7ab-fbe5-48dd-add0-42f05d3f2cfd name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.058373113Z" level=info msg="Stopped pod sandbox: 943df75e8c120cec9d37e4a8dd05aed3e8f310e67c6a2c33da550e6cdbd653f4" id=8663b4f2-8a78-4721-abda-44407668d7e3 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.156781 2588 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l_7c63c9be-d1ad-44b6-8b33-09af3ee314af/oauth-proxy/2.log" Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.157609 2588 generic.go:296] "Generic (PLEG): container finished" podID=7c63c9be-d1ad-44b6-8b33-09af3ee314af containerID="3cbc4c089943552285b7235ee6268116978902d078a993e4f88285e0c16777ce" exitCode=0 Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.157634 2588 generic.go:296] "Generic (PLEG): container finished" podID=7c63c9be-d1ad-44b6-8b33-09af3ee314af containerID="b7574d11ea7895beed8a146040a8adb91835493e426424dd8e461a93b6f7677b" exitCode=0 Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.157651 2588 generic.go:296] "Generic (PLEG): container finished" podID=7c63c9be-d1ad-44b6-8b33-09af3ee314af containerID="c507c8393ab5e7e8bb6dbc97bcabc3f2598889f9b13fbee87614f2fee17b1e9d" exitCode=0 Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.157668 2588 generic.go:296] "Generic (PLEG): container finished" podID=7c63c9be-d1ad-44b6-8b33-09af3ee314af containerID="76d0bbfe3b04b3fb31c951bf14c59dddccfd473352dd9388eeee0e7e51e96358" exitCode=2 Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.157683 2588 generic.go:296] "Generic (PLEG): container finished" podID=7c63c9be-d1ad-44b6-8b33-09af3ee314af containerID="578cce0ed1aa4795c3b6d8648ecf4b937233d871e6a027eb233bec20d4333c1f" exitCode=0 Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.157738 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" event=&{ID:7c63c9be-d1ad-44b6-8b33-09af3ee314af Type:ContainerDied Data:3cbc4c089943552285b7235ee6268116978902d078a993e4f88285e0c16777ce} Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.157762 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" event=&{ID:7c63c9be-d1ad-44b6-8b33-09af3ee314af Type:ContainerDied Data:b7574d11ea7895beed8a146040a8adb91835493e426424dd8e461a93b6f7677b} Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.157778 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" event=&{ID:7c63c9be-d1ad-44b6-8b33-09af3ee314af Type:ContainerDied Data:c507c8393ab5e7e8bb6dbc97bcabc3f2598889f9b13fbee87614f2fee17b1e9d} Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.157793 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" event=&{ID:7c63c9be-d1ad-44b6-8b33-09af3ee314af Type:ContainerDied Data:76d0bbfe3b04b3fb31c951bf14c59dddccfd473352dd9388eeee0e7e51e96358} Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.157808 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" event=&{ID:7c63c9be-d1ad-44b6-8b33-09af3ee314af Type:ContainerDied Data:578cce0ed1aa4795c3b6d8648ecf4b937233d871e6a027eb233bec20d4333c1f} Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.159344 2588 generic.go:296] "Generic (PLEG): container finished" podID=fc45e847-8f55-4c48-8c48-4bfee2419dca containerID="2952a168f0605321dd4ff798bf4481d0ede73526e359849e9e7039226b853fab" exitCode=2 Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.159368 2588 generic.go:296] "Generic (PLEG): container finished" podID=fc45e847-8f55-4c48-8c48-4bfee2419dca containerID="2c1936a19655345a892373f86a623f7409e46829cbb1adeca75e2879d268aa08" exitCode=2 Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.159416 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" event=&{ID:fc45e847-8f55-4c48-8c48-4bfee2419dca Type:ContainerDied Data:2952a168f0605321dd4ff798bf4481d0ede73526e359849e9e7039226b853fab} Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.159438 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" event=&{ID:fc45e847-8f55-4c48-8c48-4bfee2419dca Type:ContainerDied Data:2c1936a19655345a892373f86a623f7409e46829cbb1adeca75e2879d268aa08} Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.164095 2588 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_05c9a6ea-140a-49a2-bde5-f220b05aa252/alertmanager-proxy/2.log" Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.164467 2588 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_05c9a6ea-140a-49a2-bde5-f220b05aa252/config-reloader/2.log" Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.164770 2588 generic.go:296] "Generic (PLEG): container finished" podID=05c9a6ea-140a-49a2-bde5-f220b05aa252 containerID="4a95d41273862399b3cca16471b999defcccef3ceb8402d7cfb798378ffc8cab" exitCode=0 Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.164791 2588 generic.go:296] "Generic (PLEG): container finished" podID=05c9a6ea-140a-49a2-bde5-f220b05aa252 containerID="53a26d2c619d3ef8230bda385bc9245dc43ada80f94fe54f1f79b82763d2f7a8" exitCode=0 Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.164808 2588 generic.go:296] "Generic (PLEG): container finished" podID=05c9a6ea-140a-49a2-bde5-f220b05aa252 containerID="85c843961a0e8a2283e240b3c350b3d3823e13fbe3cb8f291a0e809df84275e7" exitCode=2 Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.164823 2588 generic.go:296] "Generic (PLEG): container finished" podID=05c9a6ea-140a-49a2-bde5-f220b05aa252 containerID="df204cb60308dd066ca32f1f38b7a71d29064341d03539fedabfb840583c8a96" exitCode=2 Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.164842 2588 generic.go:296] "Generic (PLEG): container finished" podID=05c9a6ea-140a-49a2-bde5-f220b05aa252 containerID="29d242ba69edfb0ca301e890f9169a7355a0889dbda9e344277448360467524b" exitCode=0 Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.164887 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event=&{ID:05c9a6ea-140a-49a2-bde5-f220b05aa252 Type:ContainerDied Data:4a95d41273862399b3cca16471b999defcccef3ceb8402d7cfb798378ffc8cab} Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.164908 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event=&{ID:05c9a6ea-140a-49a2-bde5-f220b05aa252 Type:ContainerDied Data:53a26d2c619d3ef8230bda385bc9245dc43ada80f94fe54f1f79b82763d2f7a8} Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.164924 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event=&{ID:05c9a6ea-140a-49a2-bde5-f220b05aa252 Type:ContainerDied Data:85c843961a0e8a2283e240b3c350b3d3823e13fbe3cb8f291a0e809df84275e7} Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.164940 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event=&{ID:05c9a6ea-140a-49a2-bde5-f220b05aa252 Type:ContainerDied Data:df204cb60308dd066ca32f1f38b7a71d29064341d03539fedabfb840583c8a96} Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.164956 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event=&{ID:05c9a6ea-140a-49a2-bde5-f220b05aa252 Type:ContainerDied Data:29d242ba69edfb0ca301e890f9169a7355a0889dbda9e344277448360467524b} Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.165969 2588 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator_migrator-6cd87bdff8-wj8dg_08c04948-e189-4227-bece-151bf3576721/migrator/0.log" Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.166002 2588 generic.go:296] "Generic (PLEG): container finished" podID=08c04948-e189-4227-bece-151bf3576721 containerID="ae93d6fe80cbb19038b530607e317c9c10827d02079127cc185baa8cd05993b9" exitCode=2 Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.166047 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-6cd87bdff8-wj8dg" event=&{ID:08c04948-e189-4227-bece-151bf3576721 Type:ContainerDied Data:ae93d6fe80cbb19038b530607e317c9c10827d02079127cc185baa8cd05993b9} Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.167924 2588 generic.go:296] "Generic (PLEG): container finished" podID=2ecd2e2c-ab52-42e0-8fde-64e45804cc9e containerID="987ad82f0f8b3760e232b876a53eb03f7fa455eadddd9e240e5e8be4f57bf369" exitCode=2 Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.168028 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" event=&{ID:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e Type:ContainerDied Data:987ad82f0f8b3760e232b876a53eb03f7fa455eadddd9e240e5e8be4f57bf369} Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.168672635Z" level=info msg="Stopped container a988dfd069a4d8b4e201e470b1b75848391f749fd271851262309d39a93dc410: openshift-monitoring/prometheus-k8s-0/prometheus-proxy" id=1f923910-83f5-4df2-9ce0-35766e81c5c8 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.172906759Z" level=info msg="Stopped container 10f645c424bd560dd410724c8e884a10049cfe720e59016dcb3f8d51623b3ae6: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/kube-rbac-proxy-main" id=833c265e-f5eb-4f1d-89ec-66803605a5ff name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.174484 2588 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879/prometheus-proxy/2.log" Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.179799 2588 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator_migrator-6cd87bdff8-wj8dg_08c04948-e189-4227-bece-151bf3576721/migrator/0.log" Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.181471 2588 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879/config-reloader/2.log" Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.184009 2588 generic.go:296] "Generic (PLEG): container finished" podID=4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 containerID="a988dfd069a4d8b4e201e470b1b75848391f749fd271851262309d39a93dc410" exitCode=2 Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.184033 2588 generic.go:296] "Generic (PLEG): container finished" podID=4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 containerID="7a2f8f0ad0be821f8959bfd50621160131d618f2b80fdd07c0192537a6788d3a" exitCode=0 Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.184048 2588 generic.go:296] "Generic (PLEG): container finished" podID=4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 containerID="bef5095b779c6d727ff2f590bfd7c28dbad80564fc113cb2c36ebaa3ebaeaedf" exitCode=2 Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.184061 2588 generic.go:296] "Generic (PLEG): container finished" podID=4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 containerID="4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73" exitCode=0 Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.184105 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event=&{ID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Type:ContainerDied Data:a988dfd069a4d8b4e201e470b1b75848391f749fd271851262309d39a93dc410} Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.184127 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event=&{ID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Type:ContainerDied Data:7a2f8f0ad0be821f8959bfd50621160131d618f2b80fdd07c0192537a6788d3a} Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.184164 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event=&{ID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Type:ContainerDied Data:bef5095b779c6d727ff2f590bfd7c28dbad80564fc113cb2c36ebaa3ebaeaedf} Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.184179 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event=&{ID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Type:ContainerDied Data:4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73} Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.185804 2588 generic.go:296] "Generic (PLEG): container finished" podID=88745ab7-efe7-42c0-a2e8-75591518333d containerID="0b999d4ad314ba518c86d2e8df430b23bade34adf813f2325e76c8d1d79e1b52" exitCode=0 Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.185857 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" event=&{ID:88745ab7-efe7-42c0-a2e8-75591518333d Type:ContainerDied Data:0b999d4ad314ba518c86d2e8df430b23bade34adf813f2325e76c8d1d79e1b52} Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.187893 2588 generic.go:296] "Generic (PLEG): container finished" podID=9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 containerID="0f22074c4c974c98091736b511fa5de52c01939cfdd9697e4eee429370ad1540" exitCode=2 Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.187917 2588 generic.go:296] "Generic (PLEG): container finished" podID=9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 containerID="489142621cb0989124915d3c9efa7b65c6ff3af9ed775b03b9f887e6ad7d1d8b" exitCode=0 Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.187935 2588 generic.go:296] "Generic (PLEG): container finished" podID=9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 containerID="10f645c424bd560dd410724c8e884a10049cfe720e59016dcb3f8d51623b3ae6" exitCode=0 Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.187959 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" event=&{ID:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 Type:ContainerDied Data:0f22074c4c974c98091736b511fa5de52c01939cfdd9697e4eee429370ad1540} Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.187980 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" event=&{ID:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 Type:ContainerDied Data:489142621cb0989124915d3c9efa7b65c6ff3af9ed775b03b9f887e6ad7d1d8b} Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.187997 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" event=&{ID:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 Type:ContainerDied Data:10f645c424bd560dd410724c8e884a10049cfe720e59016dcb3f8d51623b3ae6} Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.203346124Z" level=info msg="Stopped container 2952a168f0605321dd4ff798bf4481d0ede73526e359849e9e7039226b853fab: openshift-monitoring/telemeter-client-5598c645c4-mftdv/reload" id=1d534ed0-b6ef-4c25-a69f-14b99b66d67e name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.223339 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k224n\" (UniqueName: \"kubernetes.io/projected/08c04948-e189-4227-bece-151bf3576721-kube-api-access-k224n\") pod \"08c04948-e189-4227-bece-151bf3576721\" (UID: \"08c04948-e189-4227-bece-151bf3576721\") " Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.244198 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08c04948-e189-4227-bece-151bf3576721-kube-api-access-k224n" (OuterVolumeSpecName: "kube-api-access-k224n") pod "08c04948-e189-4227-bece-151bf3576721" (UID: "08c04948-e189-4227-bece-151bf3576721"). InnerVolumeSpecName "kube-api-access-k224n". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.244940382Z" level=info msg="Stopped container 7a2f8f0ad0be821f8959bfd50621160131d618f2b80fdd07c0192537a6788d3a: openshift-monitoring/prometheus-k8s-0/thanos-sidecar" id=766fa700-b41c-49cf-aea9-8c2e9370f442 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.284372537Z" level=info msg="Stopped container 85c843961a0e8a2283e240b3c350b3d3823e13fbe3cb8f291a0e809df84275e7: openshift-monitoring/alertmanager-main-0/alertmanager-proxy" id=91d89d57-8d05-4c54-81cd-a02c2316e92f name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.317811675Z" level=info msg="Stopped container bef5095b779c6d727ff2f590bfd7c28dbad80564fc113cb2c36ebaa3ebaeaedf: openshift-monitoring/prometheus-k8s-0/config-reloader" id=6bbd4497-9411-45f4-8746-e17903cae2a1 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.324727 2588 reconciler.go:399] "Volume detached for volume \"kube-api-access-k224n\" (UniqueName: \"kubernetes.io/projected/08c04948-e189-4227-bece-151bf3576721-kube-api-access-k224n\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.346891377Z" level=info msg="Stopped container 489142621cb0989124915d3c9efa7b65c6ff3af9ed775b03b9f887e6ad7d1d8b: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/kube-rbac-proxy-self" id=b51ecc2a-0b68-4a4c-8868-72865a398b62 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.349018752Z" level=info msg="Stopping pod sandbox: 7a37710cfebf59892a4d6b3c2e12765b60414f6a7cfb8933a050df609c94e690" id=fc96502e-26b5-423b-9d79-c46c9dda2f9b name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.349493122Z" level=info msg="Got pod network &{Name:openshift-state-metrics-6f5dcdf79b-wq256 Namespace:openshift-monitoring ID:7a37710cfebf59892a4d6b3c2e12765b60414f6a7cfb8933a050df609c94e690 UID:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 NetNS:/var/run/netns/e7429438-8b02-46d6-a53c-f91c31850d37 Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.349638389Z" level=info msg="Deleting pod openshift-monitoring_openshift-state-metrics-6f5dcdf79b-wq256 from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.383467045Z" level=info msg="Stopped container 29d242ba69edfb0ca301e890f9169a7355a0889dbda9e344277448360467524b: openshift-monitoring/alertmanager-main-0/alertmanager" id=95d74e1c-0308-47ba-84f8-0cd0bea52eb7 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.400579724Z" level=info msg="Stopped container d2de249b9456c83b6cb07b3c7fc5010802a974d250e6d15452f354a4100685f6: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-rbac-proxy-self" id=8676d75b-e340-4481-b963-f21b1de22117 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.426692078Z" level=info msg="Stopped container 46e4da5180913795c57fd23e927474161e2562ce8f557a4ef7d187e1da8df07c: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-rbac-proxy-main" id=3711d209-e841-491f-93ab-0c6d4dc36307 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.427581521Z" level=info msg="Stopping pod sandbox: 1e97b2e64ee6d4f76d829e9619da041f24e571e3de5454c48314d411b8770c02" id=d0c4bf3c-a009-4206-b8be-3fe6d30d037c name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.427987315Z" level=info msg="Got pod network &{Name:kube-state-metrics-6798b548b9-xl88q Namespace:openshift-monitoring ID:1e97b2e64ee6d4f76d829e9619da041f24e571e3de5454c48314d411b8770c02 UID:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e NetNS:/var/run/netns/cbe37ad3-4a92-423a-8f3c-dda00e586a03 Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.428165063Z" level=info msg="Deleting pod openshift-monitoring_kube-state-metrics-6798b548b9-xl88q from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.449017338Z" level=info msg="Stopped container 7eca8a735d65a0f0812794e42a36fd300146f8782a35f3df7830a91dfdd49bef: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy-metrics" id=b9a42aed-536b-4fe5-b0bb-ea6cdcb00b56 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.449596560Z" level=info msg="Stopping pod sandbox: fcdbc115e8986f7ba4e1fc9bfa0728fd149a52733e5f0956b91d0b8262b33ca6" id=6f51ddf8-4c59-41ec-b925-9cc371911c1b name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.450057285Z" level=info msg="Got pod network &{Name:thanos-querier-6f4f5fb7cf-rdm9l Namespace:openshift-monitoring ID:fcdbc115e8986f7ba4e1fc9bfa0728fd149a52733e5f0956b91d0b8262b33ca6 UID:7c63c9be-d1ad-44b6-8b33-09af3ee314af NetNS:/var/run/netns/d054aa63-f5ac-4893-88ff-2a7d67c9ea25 Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.450231472Z" level=info msg="Deleting pod openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.459744289Z" level=info msg="Stopped container 543f2356973a019ec622df9cf319f02abc2eac1228b813bef15ed997a118fe6d: openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh/prometheus-adapter" id=5c8a419b-b6c3-4f78-9e29-6480a77d48f7 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.461792510Z" level=info msg="Stopping pod sandbox: 9d9807a6998df2507a4aa5b9bad79e347bc60d62b254989738c927929d827e6e" id=6d7204f5-c319-43cc-99d4-b229f6d2c509 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.462196905Z" level=info msg="Got pod network &{Name:prometheus-adapter-5894cc86c7-qhmbh Namespace:openshift-monitoring ID:9d9807a6998df2507a4aa5b9bad79e347bc60d62b254989738c927929d827e6e UID:a40bf20d-67a4-45e7-b701-ca5a75376c1f NetNS:/var/run/netns/2bb8265e-9d3b-4130-ab12-f8f63e286c5d Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.462349584Z" level=info msg="Deleting pod openshift-monitoring_prometheus-adapter-5894cc86c7-qhmbh from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:44:48Z [verbose] Del: openshift-monitoring:prometheus-operator-admission-webhook-598468bb49-94dsn:88745ab7-efe7-42c0-a2e8-75591518333d:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.475789596Z" level=info msg="Stopped container 7de6261b7ebb7d69cfdcbb24569dafb351175e25ae8aecd812d15cab2e644276: openshift-monitoring/prometheus-k8s-0/kube-rbac-proxy" id=4964390b-5cd2-4351-9b78-efd30ead6a0e name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.524061855Z" level=info msg="Stopped container edb583e71afefdc5a9b0a5e0042237321ce66a47c359a57e1eec8611ac5ed37d: openshift-monitoring/prometheus-k8s-0/kube-rbac-proxy-thanos" id=d0a21dd9-1bdc-4b8c-886c-9c65d6e1083d name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.524640016Z" level=info msg="Stopping pod sandbox: 575fca5742eeb550505988169b65e39889d4e9a4a916a669f68c988db3eb4df1" id=5c36e6c8-5cf7-4e53-9089-5c5cc2eefa37 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.544409481Z" level=info msg="Stopped container 4c970eddded2809e2e506cfb0b207a19b375bc49b3a1f9bfd58b202b7c1f6ebd: openshift-monitoring/alertmanager-main-0/kube-rbac-proxy-metric" id=d5e4f5ab-0f11-4cf0-ad26-e8c0fc688e07 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.544923290Z" level=info msg="Stopping pod sandbox: d6d0c82bc9774e51559f70121d26a57c7c38cce6fb83f23b63a8bb671c408015" id=76ba6536-e76b-4217-ab96-85fbd1f0d5c1 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.545362189Z" level=info msg="Got pod network &{Name:alertmanager-main-0 Namespace:openshift-monitoring ID:d6d0c82bc9774e51559f70121d26a57c7c38cce6fb83f23b63a8bb671c408015 UID:05c9a6ea-140a-49a2-bde5-f220b05aa252 NetNS:/var/run/netns/c7a80f78-1128-4e9e-9931-f7ed057df8ad Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.545503919Z" level=info msg="Deleting pod openshift-monitoring_alertmanager-main-0 from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.545843285Z" level=info msg="Stopped container 91947bee1cff242d8ff480a4ba98a9c65dd41326ef2bfc74ccc1a969906e3b00: openshift-monitoring/telemeter-client-5598c645c4-mftdv/kube-rbac-proxy" id=96cbbb48-6558-44d6-8e6c-e906e947dba2 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.546239926Z" level=info msg="Stopping pod sandbox: f9cc29d007bd3c948c46f781abd08afdb46359ec50fc31f8c503fcdb35b7aa22" id=94a26d46-d966-46fb-8726-c7086586bb30 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.546685780Z" level=info msg="Got pod network &{Name:telemeter-client-5598c645c4-mftdv Namespace:openshift-monitoring ID:f9cc29d007bd3c948c46f781abd08afdb46359ec50fc31f8c503fcdb35b7aa22 UID:fc45e847-8f55-4c48-8c48-4bfee2419dca NetNS:/var/run/netns/cac8a001-4260-4cae-ad1e-0de0816fe057 Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.546822885Z" level=info msg="Deleting pod openshift-monitoring_telemeter-client-5598c645c4-mftdv from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.597469040Z" level=info msg="Stopped pod sandbox: e10e02d14faf7624d78743e0d1899ddf64963db130ecf653b0849bcde995a72c" id=85691975-3563-4104-a900-da620801d8b9 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:44:48Z [verbose] Del: openshift-monitoring:openshift-state-metrics-6f5dcdf79b-wq256:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.725446553Z" level=info msg="Stopped pod sandbox: 7a37710cfebf59892a4d6b3c2e12765b60414f6a7cfb8933a050df609c94e690" id=fc96502e-26b5-423b-9d79-c46c9dda2f9b name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:44:48Z [verbose] Del: openshift-monitoring:thanos-querier-6f4f5fb7cf-rdm9l:7c63c9be-d1ad-44b6-8b33-09af3ee314af:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.842216 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-openshift-state-metrics-tls\") pod \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\" (UID: \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\") " Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.842295 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-openshift-state-metrics-kube-rbac-proxy-config\") pod \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\" (UID: \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\") " Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.842328 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/88745ab7-efe7-42c0-a2e8-75591518333d-tls-certificates\") pod \"88745ab7-efe7-42c0-a2e8-75591518333d\" (UID: \"88745ab7-efe7-42c0-a2e8-75591518333d\") " Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.842382 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6g5r\" (UniqueName: \"kubernetes.io/projected/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-kube-api-access-h6g5r\") pod \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\" (UID: \"9de5de36-9b80-4a0b-a615-f0f9ccda2ff6\") " Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.866598 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-openshift-state-metrics-kube-rbac-proxy-config" (OuterVolumeSpecName: "openshift-state-metrics-kube-rbac-proxy-config") pod "9de5de36-9b80-4a0b-a615-f0f9ccda2ff6" (UID: "9de5de36-9b80-4a0b-a615-f0f9ccda2ff6"). InnerVolumeSpecName "openshift-state-metrics-kube-rbac-proxy-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.870535 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-kube-api-access-h6g5r" (OuterVolumeSpecName: "kube-api-access-h6g5r") pod "9de5de36-9b80-4a0b-a615-f0f9ccda2ff6" (UID: "9de5de36-9b80-4a0b-a615-f0f9ccda2ff6"). InnerVolumeSpecName "kube-api-access-h6g5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.870728 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88745ab7-efe7-42c0-a2e8-75591518333d-tls-certificates" (OuterVolumeSpecName: "tls-certificates") pod "88745ab7-efe7-42c0-a2e8-75591518333d" (UID: "88745ab7-efe7-42c0-a2e8-75591518333d"). InnerVolumeSpecName "tls-certificates". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.871851 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-openshift-state-metrics-tls" (OuterVolumeSpecName: "openshift-state-metrics-tls") pod "9de5de36-9b80-4a0b-a615-f0f9ccda2ff6" (UID: "9de5de36-9b80-4a0b-a615-f0f9ccda2ff6"). InnerVolumeSpecName "openshift-state-metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:48.887499881Z" level=info msg="Stopped pod sandbox: fcdbc115e8986f7ba4e1fc9bfa0728fd149a52733e5f0956b91d0b8262b33ca6" id=6f51ddf8-4c59-41ec-b925-9cc371911c1b name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.900489 2588 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l_7c63c9be-d1ad-44b6-8b33-09af3ee314af/oauth-proxy/2.log" Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.945644 2588 reconciler.go:399] "Volume detached for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-openshift-state-metrics-tls\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.945686 2588 reconciler.go:399] "Volume detached for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-openshift-state-metrics-kube-rbac-proxy-config\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.945708 2588 reconciler.go:399] "Volume detached for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/88745ab7-efe7-42c0-a2e8-75591518333d-tls-certificates\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:48.945726 2588 reconciler.go:399] "Volume detached for volume \"kube-api-access-h6g5r\" (UniqueName: \"kubernetes.io/projected/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6-kube-api-access-h6g5r\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:44:48Z [verbose] Del: openshift-monitoring:kube-state-metrics-6798b548b9-xl88q:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:49.050438050Z" level=info msg="Stopped pod sandbox: 1e97b2e64ee6d4f76d829e9619da041f24e571e3de5454c48314d411b8770c02" id=d0c4bf3c-a009-4206-b8be-3fe6d30d037c name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.059298 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/7c63c9be-d1ad-44b6-8b33-09af3ee314af-metrics-client-ca\") pod \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.059351 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.059392 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"thanos-querier-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c63c9be-d1ad-44b6-8b33-09af3ee314af-thanos-querier-trusted-ca-bundle\") pod \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.059426 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvz7l\" (UniqueName: \"kubernetes.io/projected/7c63c9be-d1ad-44b6-8b33-09af3ee314af-kube-api-access-pvz7l\") pod \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.059468 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy\") pod \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.059503 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-thanos-querier-oauth-cookie\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-oauth-cookie\") pod \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.059539 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-grpc-tls\") pod \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.059571 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-tls\") pod \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.059606 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\" (UID: \"7c63c9be-d1ad-44b6-8b33-09af3ee314af\") " Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:44:49.060483 2588 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/7c63c9be-d1ad-44b6-8b33-09af3ee314af/volumes/kubernetes.io~configmap/metrics-client-ca: clearQuota called, but quotas disabled Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.060698 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c63c9be-d1ad-44b6-8b33-09af3ee314af-metrics-client-ca" (OuterVolumeSpecName: "metrics-client-ca") pod "7c63c9be-d1ad-44b6-8b33-09af3ee314af" (UID: "7c63c9be-d1ad-44b6-8b33-09af3ee314af"). InnerVolumeSpecName "metrics-client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:44:49.061068 2588 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/7c63c9be-d1ad-44b6-8b33-09af3ee314af/volumes/kubernetes.io~configmap/thanos-querier-trusted-ca-bundle: clearQuota called, but quotas disabled Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.061350 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c63c9be-d1ad-44b6-8b33-09af3ee314af-thanos-querier-trusted-ca-bundle" (OuterVolumeSpecName: "thanos-querier-trusted-ca-bundle") pod "7c63c9be-d1ad-44b6-8b33-09af3ee314af" (UID: "7c63c9be-d1ad-44b6-8b33-09af3ee314af"). InnerVolumeSpecName "thanos-querier-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.137611 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c63c9be-d1ad-44b6-8b33-09af3ee314af-kube-api-access-pvz7l" (OuterVolumeSpecName: "kube-api-access-pvz7l") pod "7c63c9be-d1ad-44b6-8b33-09af3ee314af" (UID: "7c63c9be-d1ad-44b6-8b33-09af3ee314af"). InnerVolumeSpecName "kube-api-access-pvz7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.137819 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy-metrics" (OuterVolumeSpecName: "secret-thanos-querier-kube-rbac-proxy-metrics") pod "7c63c9be-d1ad-44b6-8b33-09af3ee314af" (UID: "7c63c9be-d1ad-44b6-8b33-09af3ee314af"). InnerVolumeSpecName "secret-thanos-querier-kube-rbac-proxy-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.156986 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-grpc-tls" (OuterVolumeSpecName: "secret-grpc-tls") pod "7c63c9be-d1ad-44b6-8b33-09af3ee314af" (UID: "7c63c9be-d1ad-44b6-8b33-09af3ee314af"). InnerVolumeSpecName "secret-grpc-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.162573 2588 reconciler.go:399] "Volume detached for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/7c63c9be-d1ad-44b6-8b33-09af3ee314af-metrics-client-ca\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.162619 2588 reconciler.go:399] "Volume detached for volume \"thanos-querier-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c63c9be-d1ad-44b6-8b33-09af3ee314af-thanos-querier-trusted-ca-bundle\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.162641 2588 reconciler.go:399] "Volume detached for volume \"kube-api-access-pvz7l\" (UniqueName: \"kubernetes.io/projected/7c63c9be-d1ad-44b6-8b33-09af3ee314af-kube-api-access-pvz7l\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.162663 2588 reconciler.go:399] "Volume detached for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-grpc-tls\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.162690 2588 reconciler.go:399] "Volume detached for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy-metrics\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.165426 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy" (OuterVolumeSpecName: "secret-thanos-querier-kube-rbac-proxy") pod "7c63c9be-d1ad-44b6-8b33-09af3ee314af" (UID: "7c63c9be-d1ad-44b6-8b33-09af3ee314af"). InnerVolumeSpecName "secret-thanos-querier-kube-rbac-proxy". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.169503 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy-rules" (OuterVolumeSpecName: "secret-thanos-querier-kube-rbac-proxy-rules") pod "7c63c9be-d1ad-44b6-8b33-09af3ee314af" (UID: "7c63c9be-d1ad-44b6-8b33-09af3ee314af"). InnerVolumeSpecName "secret-thanos-querier-kube-rbac-proxy-rules". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.177458 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-tls" (OuterVolumeSpecName: "secret-thanos-querier-tls") pod "7c63c9be-d1ad-44b6-8b33-09af3ee314af" (UID: "7c63c9be-d1ad-44b6-8b33-09af3ee314af"). InnerVolumeSpecName "secret-thanos-querier-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:44:48Z [verbose] Del: openshift-monitoring:prometheus-adapter-5894cc86c7-qhmbh:a40bf20d-67a4-45e7-b701-ca5a75376c1f:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.197375 2588 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_05c9a6ea-140a-49a2-bde5-f220b05aa252/alertmanager-proxy/2.log" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.197806 2588 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_05c9a6ea-140a-49a2-bde5-f220b05aa252/config-reloader/2.log" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.198241 2588 generic.go:296] "Generic (PLEG): container finished" podID=05c9a6ea-140a-49a2-bde5-f220b05aa252 containerID="4c970eddded2809e2e506cfb0b207a19b375bc49b3a1f9bfd58b202b7c1f6ebd" exitCode=0 Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.198305 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event=&{ID:05c9a6ea-140a-49a2-bde5-f220b05aa252 Type:ContainerDied Data:4c970eddded2809e2e506cfb0b207a19b375bc49b3a1f9bfd58b202b7c1f6ebd} Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.199856 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256" event=&{ID:9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 Type:ContainerDied Data:7a37710cfebf59892a4d6b3c2e12765b60414f6a7cfb8933a050df609c94e690} Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.199886 2588 scope.go:115] "RemoveContainer" containerID="0f22074c4c974c98091736b511fa5de52c01939cfdd9697e4eee429370ad1540" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.207731 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-oauth-cookie" (OuterVolumeSpecName: "secret-thanos-querier-oauth-cookie") pod "7c63c9be-d1ad-44b6-8b33-09af3ee314af" (UID: "7c63c9be-d1ad-44b6-8b33-09af3ee314af"). InnerVolumeSpecName "secret-thanos-querier-oauth-cookie". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:49.210216660Z" level=info msg="Removing container: 0f22074c4c974c98091736b511fa5de52c01939cfdd9697e4eee429370ad1540" id=423f7e94-dcbb-4d93-a9a9-08b6e14121e5 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.215439 2588 generic.go:296] "Generic (PLEG): container finished" podID=a40bf20d-67a4-45e7-b701-ca5a75376c1f containerID="543f2356973a019ec622df9cf319f02abc2eac1228b813bef15ed997a118fe6d" exitCode=0 Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.215495 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" event=&{ID:a40bf20d-67a4-45e7-b701-ca5a75376c1f Type:ContainerDied Data:543f2356973a019ec622df9cf319f02abc2eac1228b813bef15ed997a118fe6d} Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.235404 2588 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256] Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.235795 2588 generic.go:296] "Generic (PLEG): container finished" podID=fc45e847-8f55-4c48-8c48-4bfee2419dca containerID="91947bee1cff242d8ff480a4ba98a9c65dd41326ef2bfc74ccc1a969906e3b00" exitCode=0 Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.235845 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" event=&{ID:fc45e847-8f55-4c48-8c48-4bfee2419dca Type:ContainerDied Data:91947bee1cff242d8ff480a4ba98a9c65dd41326ef2bfc74ccc1a969906e3b00} Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.236960 2588 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-kube-storage-version-migrator_migrator-6cd87bdff8-wj8dg_08c04948-e189-4227-bece-151bf3576721/migrator/0.log" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.237020 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-6cd87bdff8-wj8dg" event=&{ID:08c04948-e189-4227-bece-151bf3576721 Type:ContainerDied Data:943df75e8c120cec9d37e4a8dd05aed3e8f310e67c6a2c33da550e6cdbd653f4} Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.240375 2588 generic.go:296] "Generic (PLEG): container finished" podID=2ecd2e2c-ab52-42e0-8fde-64e45804cc9e containerID="d2de249b9456c83b6cb07b3c7fc5010802a974d250e6d15452f354a4100685f6" exitCode=0 Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.240397 2588 generic.go:296] "Generic (PLEG): container finished" podID=2ecd2e2c-ab52-42e0-8fde-64e45804cc9e containerID="46e4da5180913795c57fd23e927474161e2562ce8f557a4ef7d187e1da8df07c" exitCode=0 Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.240488 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" event=&{ID:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e Type:ContainerDied Data:d2de249b9456c83b6cb07b3c7fc5010802a974d250e6d15452f354a4100685f6} Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.240510 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" event=&{ID:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e Type:ContainerDied Data:46e4da5180913795c57fd23e927474161e2562ce8f557a4ef7d187e1da8df07c} Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.240538 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-6798b548b9-xl88q" event=&{ID:2ecd2e2c-ab52-42e0-8fde-64e45804cc9e Type:ContainerDied Data:1e97b2e64ee6d4f76d829e9619da041f24e571e3de5454c48314d411b8770c02} Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.242670 2588 kubelet.go:2129] "SyncLoop REMOVE" source="api" pods=[openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256] Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.268776 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-metrics-client-ca\") pod \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.268840 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blpwr\" (UniqueName: \"kubernetes.io/projected/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-api-access-blpwr\") pod \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.268889 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-state-metrics-tls\") pod \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.268928 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-state-metrics-kube-rbac-proxy-config\") pod \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.268978 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-volume-directive-shadow\") pod \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\" (UID: \"2ecd2e2c-ab52-42e0-8fde-64e45804cc9e\") " Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.269206 2588 reconciler.go:399] "Volume detached for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-tls\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.269230 2588 reconciler.go:399] "Volume detached for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy-rules\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.269254 2588 reconciler.go:399] "Volume detached for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-kube-rbac-proxy\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.269285 2588 reconciler.go:399] "Volume detached for volume \"secret-thanos-querier-oauth-cookie\" (UniqueName: \"kubernetes.io/secret/7c63c9be-d1ad-44b6-8b33-09af3ee314af-secret-thanos-querier-oauth-cookie\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:44:49.269459 2588 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e/volumes/kubernetes.io~empty-dir/volume-directive-shadow: clearQuota called, but quotas disabled Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.269499 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-volume-directive-shadow" (OuterVolumeSpecName: "volume-directive-shadow") pod "2ecd2e2c-ab52-42e0-8fde-64e45804cc9e" (UID: "2ecd2e2c-ab52-42e0-8fde-64e45804cc9e"). InnerVolumeSpecName "volume-directive-shadow". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.300167 2588 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-kube-storage-version-migrator/migrator-6cd87bdff8-wj8dg] Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.300219 2588 kubelet.go:2129] "SyncLoop REMOVE" source="api" pods=[openshift-kube-storage-version-migrator/migrator-6cd87bdff8-wj8dg] Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:44:49.310985 2588 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e/volumes/kubernetes.io~configmap/metrics-client-ca: clearQuota called, but quotas disabled Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.311200 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-metrics-client-ca" (OuterVolumeSpecName: "metrics-client-ca") pod "2ecd2e2c-ab52-42e0-8fde-64e45804cc9e" (UID: "2ecd2e2c-ab52-42e0-8fde-64e45804cc9e"). InnerVolumeSpecName "metrics-client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:49.325236535Z" level=info msg="Stopped pod sandbox: 9d9807a6998df2507a4aa5b9bad79e347bc60d62b254989738c927929d827e6e" id=6d7204f5-c319-43cc-99d4-b229f6d2c509 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.326672 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-api-access-blpwr" (OuterVolumeSpecName: "kube-api-access-blpwr") pod "2ecd2e2c-ab52-42e0-8fde-64e45804cc9e" (UID: "2ecd2e2c-ab52-42e0-8fde-64e45804cc9e"). InnerVolumeSpecName "kube-api-access-blpwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.326918 2588 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879/prometheus-proxy/2.log" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.329088 2588 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879/config-reloader/2.log" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.329734 2588 generic.go:296] "Generic (PLEG): container finished" podID=4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 containerID="edb583e71afefdc5a9b0a5e0042237321ce66a47c359a57e1eec8611ac5ed37d" exitCode=0 Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.329756 2588 generic.go:296] "Generic (PLEG): container finished" podID=4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 containerID="7de6261b7ebb7d69cfdcbb24569dafb351175e25ae8aecd812d15cab2e644276" exitCode=0 Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.329802 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event=&{ID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Type:ContainerDied Data:edb583e71afefdc5a9b0a5e0042237321ce66a47c359a57e1eec8611ac5ed37d} Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.329824 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event=&{ID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Type:ContainerDied Data:7de6261b7ebb7d69cfdcbb24569dafb351175e25ae8aecd812d15cab2e644276} Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.332932 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn" event=&{ID:88745ab7-efe7-42c0-a2e8-75591518333d Type:ContainerDied Data:e10e02d14faf7624d78743e0d1899ddf64963db130ecf653b0849bcde995a72c} Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:49.342907038Z" level=info msg="Removed container 0f22074c4c974c98091736b511fa5de52c01939cfdd9697e4eee429370ad1540: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/openshift-state-metrics" id=423f7e94-dcbb-4d93-a9a9-08b6e14121e5 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.343327 2588 scope.go:115] "RemoveContainer" containerID="489142621cb0989124915d3c9efa7b65c6ff3af9ed775b03b9f887e6ad7d1d8b" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.344258 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-state-metrics-kube-rbac-proxy-config" (OuterVolumeSpecName: "kube-state-metrics-kube-rbac-proxy-config") pod "2ecd2e2c-ab52-42e0-8fde-64e45804cc9e" (UID: "2ecd2e2c-ab52-42e0-8fde-64e45804cc9e"). InnerVolumeSpecName "kube-state-metrics-kube-rbac-proxy-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:49.346661429Z" level=info msg="Removing container: 489142621cb0989124915d3c9efa7b65c6ff3af9ed775b03b9f887e6ad7d1d8b" id=020a4adf-80cb-43ca-a707-7e437965e1eb name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.349821 2588 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-6f4f5fb7cf-rdm9l_7c63c9be-d1ad-44b6-8b33-09af3ee314af/oauth-proxy/2.log" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.350261 2588 generic.go:296] "Generic (PLEG): container finished" podID=7c63c9be-d1ad-44b6-8b33-09af3ee314af containerID="7eca8a735d65a0f0812794e42a36fd300146f8782a35f3df7830a91dfdd49bef" exitCode=0 Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.350314 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" event=&{ID:7c63c9be-d1ad-44b6-8b33-09af3ee314af Type:ContainerDied Data:7eca8a735d65a0f0812794e42a36fd300146f8782a35f3df7830a91dfdd49bef} Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.350338 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l" event=&{ID:7c63c9be-d1ad-44b6-8b33-09af3ee314af Type:ContainerDied Data:fcdbc115e8986f7ba4e1fc9bfa0728fd149a52733e5f0956b91d0b8262b33ca6} Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.356948 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-state-metrics-tls" (OuterVolumeSpecName: "kube-state-metrics-tls") pod "2ecd2e2c-ab52-42e0-8fde-64e45804cc9e" (UID: "2ecd2e2c-ab52-42e0-8fde-64e45804cc9e"). InnerVolumeSpecName "kube-state-metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.376577 2588 reconciler.go:399] "Volume detached for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-state-metrics-kube-rbac-proxy-config\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.376615 2588 reconciler.go:399] "Volume detached for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-volume-directive-shadow\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.376637 2588 reconciler.go:399] "Volume detached for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-metrics-client-ca\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.376656 2588 reconciler.go:399] "Volume detached for volume \"kube-api-access-blpwr\" (UniqueName: \"kubernetes.io/projected/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-api-access-blpwr\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.376675 2588 reconciler.go:399] "Volume detached for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e-kube-state-metrics-tls\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.446070 2588 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn] Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.446115 2588 kubelet.go:2129] "SyncLoop REMOVE" source="api" pods=[openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn] Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:49.476033406Z" level=info msg="Removed container 489142621cb0989124915d3c9efa7b65c6ff3af9ed775b03b9f887e6ad7d1d8b: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/kube-rbac-proxy-self" id=020a4adf-80cb-43ca-a707-7e437965e1eb name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.500673 2588 scope.go:115] "RemoveContainer" containerID="10f645c424bd560dd410724c8e884a10049cfe720e59016dcb3f8d51623b3ae6" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.501147 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"prometheus-adapter-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-prometheus-adapter-audit-profiles\") pod \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.501186 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/a40bf20d-67a4-45e7-b701-ca5a75376c1f-audit-log\") pod \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.501229 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a40bf20d-67a4-45e7-b701-ca5a75376c1f-tmpfs\") pod \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.501259 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"tls\" (UniqueName: \"kubernetes.io/secret/a40bf20d-67a4-45e7-b701-ca5a75376c1f-tls\") pod \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.501302 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-serving-certs-ca-bundle\") pod \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.501334 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfs6g\" (UniqueName: \"kubernetes.io/projected/a40bf20d-67a4-45e7-b701-ca5a75376c1f-kube-api-access-gfs6g\") pod \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.501369 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"prometheus-adapter-prometheus-config\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-prometheus-adapter-prometheus-config\") pod \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.501399 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-config\") pod \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\" (UID: \"a40bf20d-67a4-45e7-b701-ca5a75376c1f\") " Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:44:49.501664 2588 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/a40bf20d-67a4-45e7-b701-ca5a75376c1f/volumes/kubernetes.io~configmap/config: clearQuota called, but quotas disabled Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.501860 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-config" (OuterVolumeSpecName: "config") pod "a40bf20d-67a4-45e7-b701-ca5a75376c1f" (UID: "a40bf20d-67a4-45e7-b701-ca5a75376c1f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:44:49.501991 2588 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/a40bf20d-67a4-45e7-b701-ca5a75376c1f/volumes/kubernetes.io~configmap/prometheus-adapter-audit-profiles: clearQuota called, but quotas disabled Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.502317 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-prometheus-adapter-audit-profiles" (OuterVolumeSpecName: "prometheus-adapter-audit-profiles") pod "a40bf20d-67a4-45e7-b701-ca5a75376c1f" (UID: "a40bf20d-67a4-45e7-b701-ca5a75376c1f"). InnerVolumeSpecName "prometheus-adapter-audit-profiles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:44:49.502444 2588 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/a40bf20d-67a4-45e7-b701-ca5a75376c1f/volumes/kubernetes.io~empty-dir/audit-log: clearQuota called, but quotas disabled Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.502561 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a40bf20d-67a4-45e7-b701-ca5a75376c1f-audit-log" (OuterVolumeSpecName: "audit-log") pod "a40bf20d-67a4-45e7-b701-ca5a75376c1f" (UID: "a40bf20d-67a4-45e7-b701-ca5a75376c1f"). InnerVolumeSpecName "audit-log". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:44:49.502675 2588 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/a40bf20d-67a4-45e7-b701-ca5a75376c1f/volumes/kubernetes.io~empty-dir/tmpfs: clearQuota called, but quotas disabled Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.502707 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a40bf20d-67a4-45e7-b701-ca5a75376c1f-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "a40bf20d-67a4-45e7-b701-ca5a75376c1f" (UID: "a40bf20d-67a4-45e7-b701-ca5a75376c1f"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:44:49.503284 2588 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/a40bf20d-67a4-45e7-b701-ca5a75376c1f/volumes/kubernetes.io~configmap/serving-certs-ca-bundle: clearQuota called, but quotas disabled Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.503461 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-serving-certs-ca-bundle" (OuterVolumeSpecName: "serving-certs-ca-bundle") pod "a40bf20d-67a4-45e7-b701-ca5a75376c1f" (UID: "a40bf20d-67a4-45e7-b701-ca5a75376c1f"). InnerVolumeSpecName "serving-certs-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:44:49.503589 2588 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/a40bf20d-67a4-45e7-b701-ca5a75376c1f/volumes/kubernetes.io~configmap/prometheus-adapter-prometheus-config: clearQuota called, but quotas disabled Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:49.507276827Z" level=info msg="Removing container: 10f645c424bd560dd410724c8e884a10049cfe720e59016dcb3f8d51623b3ae6" id=499daa45-7dd8-4894-b832-4ab552532f3e name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.503754 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-prometheus-adapter-prometheus-config" (OuterVolumeSpecName: "prometheus-adapter-prometheus-config") pod "a40bf20d-67a4-45e7-b701-ca5a75376c1f" (UID: "a40bf20d-67a4-45e7-b701-ca5a75376c1f"). InnerVolumeSpecName "prometheus-adapter-prometheus-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.516974 2588 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l] Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.518435 2588 kubelet.go:2129] "SyncLoop REMOVE" source="api" pods=[openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l] Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.580703 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a40bf20d-67a4-45e7-b701-ca5a75376c1f-kube-api-access-gfs6g" (OuterVolumeSpecName: "kube-api-access-gfs6g") pod "a40bf20d-67a4-45e7-b701-ca5a75376c1f" (UID: "a40bf20d-67a4-45e7-b701-ca5a75376c1f"). InnerVolumeSpecName "kube-api-access-gfs6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.587728 2588 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-monitoring/kube-state-metrics-6798b548b9-xl88q] Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.590441 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a40bf20d-67a4-45e7-b701-ca5a75376c1f-tls" (OuterVolumeSpecName: "tls") pod "a40bf20d-67a4-45e7-b701-ca5a75376c1f" (UID: "a40bf20d-67a4-45e7-b701-ca5a75376c1f"). InnerVolumeSpecName "tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:49.592661850Z" level=info msg="Removed container 10f645c424bd560dd410724c8e884a10049cfe720e59016dcb3f8d51623b3ae6: openshift-monitoring/openshift-state-metrics-6f5dcdf79b-wq256/kube-rbac-proxy-main" id=499daa45-7dd8-4894-b832-4ab552532f3e name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.592970 2588 scope.go:115] "RemoveContainer" containerID="ae93d6fe80cbb19038b530607e317c9c10827d02079127cc185baa8cd05993b9" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.595316 2588 kubelet.go:2129] "SyncLoop REMOVE" source="api" pods=[openshift-monitoring/kube-state-metrics-6798b548b9-xl88q] Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:49.595822123Z" level=info msg="Removing container: ae93d6fe80cbb19038b530607e317c9c10827d02079127cc185baa8cd05993b9" id=b5eab153-b0ad-4163-8877-83acc7ffb57c name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.607089 2588 reconciler.go:399] "Volume detached for volume \"prometheus-adapter-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-prometheus-adapter-audit-profiles\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.607121 2588 reconciler.go:399] "Volume detached for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/a40bf20d-67a4-45e7-b701-ca5a75376c1f-audit-log\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.607162 2588 reconciler.go:399] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a40bf20d-67a4-45e7-b701-ca5a75376c1f-tmpfs\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.607181 2588 reconciler.go:399] "Volume detached for volume \"tls\" (UniqueName: \"kubernetes.io/secret/a40bf20d-67a4-45e7-b701-ca5a75376c1f-tls\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.607201 2588 reconciler.go:399] "Volume detached for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-serving-certs-ca-bundle\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.607221 2588 reconciler.go:399] "Volume detached for volume \"kube-api-access-gfs6g\" (UniqueName: \"kubernetes.io/projected/a40bf20d-67a4-45e7-b701-ca5a75376c1f-kube-api-access-gfs6g\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.607241 2588 reconciler.go:399] "Volume detached for volume \"prometheus-adapter-prometheus-config\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-prometheus-adapter-prometheus-config\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.607272 2588 reconciler.go:399] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a40bf20d-67a4-45e7-b701-ca5a75376c1f-config\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:44:48Z [verbose] Del: openshift-monitoring:alertmanager-main-0:05c9a6ea-140a-49a2-bde5-f220b05aa252:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:49.685298194Z" level=info msg="Removed container ae93d6fe80cbb19038b530607e317c9c10827d02079127cc185baa8cd05993b9: openshift-kube-storage-version-migrator/migrator-6cd87bdff8-wj8dg/migrator" id=b5eab153-b0ad-4163-8877-83acc7ffb57c name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.687399 2588 scope.go:115] "RemoveContainer" containerID="d2de249b9456c83b6cb07b3c7fc5010802a974d250e6d15452f354a4100685f6" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:49.688727811Z" level=info msg="Removing container: d2de249b9456c83b6cb07b3c7fc5010802a974d250e6d15452f354a4100685f6" id=1d30c19d-367a-42d7-9a01-249a3ce05dbe name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:49.858505371Z" level=info msg="Stopped pod sandbox: d6d0c82bc9774e51559f70121d26a57c7c38cce6fb83f23b63a8bb671c408015" id=76ba6536-e76b-4217-ab96-85fbd1f0d5c1 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:49.902024598Z" level=info msg="Removed container d2de249b9456c83b6cb07b3c7fc5010802a974d250e6d15452f354a4100685f6: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-rbac-proxy-self" id=1d30c19d-367a-42d7-9a01-249a3ce05dbe name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.902360 2588 scope.go:115] "RemoveContainer" containerID="46e4da5180913795c57fd23e927474161e2562ce8f557a4ef7d187e1da8df07c" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:49.903991873Z" level=info msg="Removing container: 46e4da5180913795c57fd23e927474161e2562ce8f557a4ef7d187e1da8df07c" id=50cdbb46-f646-4a21-ac14-400765996255 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.914001 2588 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_05c9a6ea-140a-49a2-bde5-f220b05aa252/alertmanager-proxy/2.log" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.919115 2588 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_05c9a6ea-140a-49a2-bde5-f220b05aa252/config-reloader/2.log" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:49.942309152Z" level=info msg="Removed container 46e4da5180913795c57fd23e927474161e2562ce8f557a4ef7d187e1da8df07c: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-rbac-proxy-main" id=50cdbb46-f646-4a21-ac14-400765996255 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:49.942903 2588 scope.go:115] "RemoveContainer" containerID="987ad82f0f8b3760e232b876a53eb03f7fa455eadddd9e240e5e8be4f57bf369" Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:49.944115564Z" level=info msg="Removing container: 987ad82f0f8b3760e232b876a53eb03f7fa455eadddd9e240e5e8be4f57bf369" id=b0a62d5a-e879-4441-a7ff-fae27d1f64de name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:44:48Z [verbose] Del: openshift-monitoring:telemeter-client-5598c645c4-mftdv:fc45e847-8f55-4c48-8c48-4bfee2419dca:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.007208712Z" level=info msg="Removed container 987ad82f0f8b3760e232b876a53eb03f7fa455eadddd9e240e5e8be4f57bf369: openshift-monitoring/kube-state-metrics-6798b548b9-xl88q/kube-state-metrics" id=b0a62d5a-e879-4441-a7ff-fae27d1f64de name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.009768182Z" level=info msg="Removing container: 0b999d4ad314ba518c86d2e8df430b23bade34adf813f2325e76c8d1d79e1b52" id=b56395fe-153d-4729-af0e-0cdd932f0e18 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.007444 2588 scope.go:115] "RemoveContainer" containerID="d2de249b9456c83b6cb07b3c7fc5010802a974d250e6d15452f354a4100685f6" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:44:50.007768 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2de249b9456c83b6cb07b3c7fc5010802a974d250e6d15452f354a4100685f6\": container with ID starting with d2de249b9456c83b6cb07b3c7fc5010802a974d250e6d15452f354a4100685f6 not found: ID does not exist" containerID="d2de249b9456c83b6cb07b3c7fc5010802a974d250e6d15452f354a4100685f6" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.007805 2588 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:d2de249b9456c83b6cb07b3c7fc5010802a974d250e6d15452f354a4100685f6} err="failed to get container status \"d2de249b9456c83b6cb07b3c7fc5010802a974d250e6d15452f354a4100685f6\": rpc error: code = NotFound desc = could not find container \"d2de249b9456c83b6cb07b3c7fc5010802a974d250e6d15452f354a4100685f6\": container with ID starting with d2de249b9456c83b6cb07b3c7fc5010802a974d250e6d15452f354a4100685f6 not found: ID does not exist" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.007816 2588 scope.go:115] "RemoveContainer" containerID="46e4da5180913795c57fd23e927474161e2562ce8f557a4ef7d187e1da8df07c" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:44:50.008041 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46e4da5180913795c57fd23e927474161e2562ce8f557a4ef7d187e1da8df07c\": container with ID starting with 46e4da5180913795c57fd23e927474161e2562ce8f557a4ef7d187e1da8df07c not found: ID does not exist" containerID="46e4da5180913795c57fd23e927474161e2562ce8f557a4ef7d187e1da8df07c" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.008075 2588 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:46e4da5180913795c57fd23e927474161e2562ce8f557a4ef7d187e1da8df07c} err="failed to get container status \"46e4da5180913795c57fd23e927474161e2562ce8f557a4ef7d187e1da8df07c\": rpc error: code = NotFound desc = could not find container \"46e4da5180913795c57fd23e927474161e2562ce8f557a4ef7d187e1da8df07c\": container with ID starting with 46e4da5180913795c57fd23e927474161e2562ce8f557a4ef7d187e1da8df07c not found: ID does not exist" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.008090 2588 scope.go:115] "RemoveContainer" containerID="987ad82f0f8b3760e232b876a53eb03f7fa455eadddd9e240e5e8be4f57bf369" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:44:50.008339 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"987ad82f0f8b3760e232b876a53eb03f7fa455eadddd9e240e5e8be4f57bf369\": container with ID starting with 987ad82f0f8b3760e232b876a53eb03f7fa455eadddd9e240e5e8be4f57bf369 not found: ID does not exist" containerID="987ad82f0f8b3760e232b876a53eb03f7fa455eadddd9e240e5e8be4f57bf369" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.008366 2588 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:987ad82f0f8b3760e232b876a53eb03f7fa455eadddd9e240e5e8be4f57bf369} err="failed to get container status \"987ad82f0f8b3760e232b876a53eb03f7fa455eadddd9e240e5e8be4f57bf369\": rpc error: code = NotFound desc = could not find container \"987ad82f0f8b3760e232b876a53eb03f7fa455eadddd9e240e5e8be4f57bf369\": container with ID starting with 987ad82f0f8b3760e232b876a53eb03f7fa455eadddd9e240e5e8be4f57bf369 not found: ID does not exist" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.008377 2588 scope.go:115] "RemoveContainer" containerID="d2de249b9456c83b6cb07b3c7fc5010802a974d250e6d15452f354a4100685f6" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.008583 2588 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:d2de249b9456c83b6cb07b3c7fc5010802a974d250e6d15452f354a4100685f6} err="failed to get container status \"d2de249b9456c83b6cb07b3c7fc5010802a974d250e6d15452f354a4100685f6\": rpc error: code = NotFound desc = could not find container \"d2de249b9456c83b6cb07b3c7fc5010802a974d250e6d15452f354a4100685f6\": container with ID starting with d2de249b9456c83b6cb07b3c7fc5010802a974d250e6d15452f354a4100685f6 not found: ID does not exist" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.008604 2588 scope.go:115] "RemoveContainer" containerID="46e4da5180913795c57fd23e927474161e2562ce8f557a4ef7d187e1da8df07c" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.008805 2588 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:46e4da5180913795c57fd23e927474161e2562ce8f557a4ef7d187e1da8df07c} err="failed to get container status \"46e4da5180913795c57fd23e927474161e2562ce8f557a4ef7d187e1da8df07c\": rpc error: code = NotFound desc = could not find container \"46e4da5180913795c57fd23e927474161e2562ce8f557a4ef7d187e1da8df07c\": container with ID starting with 46e4da5180913795c57fd23e927474161e2562ce8f557a4ef7d187e1da8df07c not found: ID does not exist" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.008825 2588 scope.go:115] "RemoveContainer" containerID="987ad82f0f8b3760e232b876a53eb03f7fa455eadddd9e240e5e8be4f57bf369" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.009022 2588 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:987ad82f0f8b3760e232b876a53eb03f7fa455eadddd9e240e5e8be4f57bf369} err="failed to get container status \"987ad82f0f8b3760e232b876a53eb03f7fa455eadddd9e240e5e8be4f57bf369\": rpc error: code = NotFound desc = could not find container \"987ad82f0f8b3760e232b876a53eb03f7fa455eadddd9e240e5e8be4f57bf369\": container with ID starting with 987ad82f0f8b3760e232b876a53eb03f7fa455eadddd9e240e5e8be4f57bf369 not found: ID does not exist" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.009043 2588 scope.go:115] "RemoveContainer" containerID="0b999d4ad314ba518c86d2e8df430b23bade34adf813f2325e76c8d1d79e1b52" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.058499277Z" level=info msg="Stopped pod sandbox: f9cc29d007bd3c948c46f781abd08afdb46359ec50fc31f8c503fcdb35b7aa22" id=94a26d46-d966-46fb-8726-c7086586bb30 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.058749 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/05c9a6ea-140a-49a2-bde5-f220b05aa252-metrics-client-ca\") pod \"05c9a6ea-140a-49a2-bde5-f220b05aa252\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.058806 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/05c9a6ea-140a-49a2-bde5-f220b05aa252-alertmanager-main-db\") pod \"05c9a6ea-140a-49a2-bde5-f220b05aa252\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.058843 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-kube-rbac-proxy-metric\") pod \"05c9a6ea-140a-49a2-bde5-f220b05aa252\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.058880 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/05c9a6ea-140a-49a2-bde5-f220b05aa252-tls-assets\") pod \"05c9a6ea-140a-49a2-bde5-f220b05aa252\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.058911 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05c9a6ea-140a-49a2-bde5-f220b05aa252-alertmanager-trusted-ca-bundle\") pod \"05c9a6ea-140a-49a2-bde5-f220b05aa252\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.058946 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-config-volume\") pod \"05c9a6ea-140a-49a2-bde5-f220b05aa252\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.058978 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/05c9a6ea-140a-49a2-bde5-f220b05aa252-config-out\") pod \"05c9a6ea-140a-49a2-bde5-f220b05aa252\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.059015 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-web-config\") pod \"05c9a6ea-140a-49a2-bde5-f220b05aa252\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.059050 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-kube-rbac-proxy\") pod \"05c9a6ea-140a-49a2-bde5-f220b05aa252\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.059092 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-alertmanager-main-proxy\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-main-proxy\") pod \"05c9a6ea-140a-49a2-bde5-f220b05aa252\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.059191 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-main-tls\") pod \"05c9a6ea-140a-49a2-bde5-f220b05aa252\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.059225 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxsm5\" (UniqueName: \"kubernetes.io/projected/05c9a6ea-140a-49a2-bde5-f220b05aa252-kube-api-access-kxsm5\") pod \"05c9a6ea-140a-49a2-bde5-f220b05aa252\" (UID: \"05c9a6ea-140a-49a2-bde5-f220b05aa252\") " Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:44:50.061406 2588 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/05c9a6ea-140a-49a2-bde5-f220b05aa252/volumes/kubernetes.io~configmap/metrics-client-ca: clearQuota called, but quotas disabled Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.061620 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05c9a6ea-140a-49a2-bde5-f220b05aa252-metrics-client-ca" (OuterVolumeSpecName: "metrics-client-ca") pod "05c9a6ea-140a-49a2-bde5-f220b05aa252" (UID: "05c9a6ea-140a-49a2-bde5-f220b05aa252"). InnerVolumeSpecName "metrics-client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:44:50.061773 2588 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/05c9a6ea-140a-49a2-bde5-f220b05aa252/volumes/kubernetes.io~empty-dir/alertmanager-main-db: clearQuota called, but quotas disabled Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.061874 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05c9a6ea-140a-49a2-bde5-f220b05aa252-alertmanager-main-db" (OuterVolumeSpecName: "alertmanager-main-db") pod "05c9a6ea-140a-49a2-bde5-f220b05aa252" (UID: "05c9a6ea-140a-49a2-bde5-f220b05aa252"). InnerVolumeSpecName "alertmanager-main-db". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:44:50.062520 2588 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/05c9a6ea-140a-49a2-bde5-f220b05aa252/volumes/kubernetes.io~configmap/alertmanager-trusted-ca-bundle: clearQuota called, but quotas disabled Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.062737 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05c9a6ea-140a-49a2-bde5-f220b05aa252-alertmanager-trusted-ca-bundle" (OuterVolumeSpecName: "alertmanager-trusted-ca-bundle") pod "05c9a6ea-140a-49a2-bde5-f220b05aa252" (UID: "05c9a6ea-140a-49a2-bde5-f220b05aa252"). InnerVolumeSpecName "alertmanager-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:44:50.063051 2588 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/05c9a6ea-140a-49a2-bde5-f220b05aa252/volumes/kubernetes.io~empty-dir/config-out: clearQuota called, but quotas disabled Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.063209 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05c9a6ea-140a-49a2-bde5-f220b05aa252-config-out" (OuterVolumeSpecName: "config-out") pod "05c9a6ea-140a-49a2-bde5-f220b05aa252" (UID: "05c9a6ea-140a-49a2-bde5-f220b05aa252"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.096462 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-config-volume" (OuterVolumeSpecName: "config-volume") pod "05c9a6ea-140a-49a2-bde5-f220b05aa252" (UID: "05c9a6ea-140a-49a2-bde5-f220b05aa252"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.097765 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05c9a6ea-140a-49a2-bde5-f220b05aa252-kube-api-access-kxsm5" (OuterVolumeSpecName: "kube-api-access-kxsm5") pod "05c9a6ea-140a-49a2-bde5-f220b05aa252" (UID: "05c9a6ea-140a-49a2-bde5-f220b05aa252"). InnerVolumeSpecName "kube-api-access-kxsm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.098397645Z" level=info msg="Removed container 0b999d4ad314ba518c86d2e8df430b23bade34adf813f2325e76c8d1d79e1b52: openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-94dsn/prometheus-operator-admission-webhook" id=b56395fe-153d-4729-af0e-0cdd932f0e18 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.098675 2588 scope.go:115] "RemoveContainer" containerID="7eca8a735d65a0f0812794e42a36fd300146f8782a35f3df7830a91dfdd49bef" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.100294240Z" level=info msg="Removing container: 7eca8a735d65a0f0812794e42a36fd300146f8782a35f3df7830a91dfdd49bef" id=07ea131a-a8db-4107-ad36-b37e78bb41af name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.121758 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05c9a6ea-140a-49a2-bde5-f220b05aa252-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "05c9a6ea-140a-49a2-bde5-f220b05aa252" (UID: "05c9a6ea-140a-49a2-bde5-f220b05aa252"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.121810 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-kube-rbac-proxy" (OuterVolumeSpecName: "secret-alertmanager-kube-rbac-proxy") pod "05c9a6ea-140a-49a2-bde5-f220b05aa252" (UID: "05c9a6ea-140a-49a2-bde5-f220b05aa252"). InnerVolumeSpecName "secret-alertmanager-kube-rbac-proxy". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.126954853Z" level=info msg="Removed container 7eca8a735d65a0f0812794e42a36fd300146f8782a35f3df7830a91dfdd49bef: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy-metrics" id=07ea131a-a8db-4107-ad36-b37e78bb41af name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.127112 2588 scope.go:115] "RemoveContainer" containerID="3cbc4c089943552285b7235ee6268116978902d078a993e4f88285e0c16777ce" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.128023957Z" level=info msg="Removing container: 3cbc4c089943552285b7235ee6268116978902d078a993e4f88285e0c16777ce" id=dd679a59-d8e8-4b2a-aba6-826719daa3ee name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.134279 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-kube-rbac-proxy-metric" (OuterVolumeSpecName: "secret-alertmanager-kube-rbac-proxy-metric") pod "05c9a6ea-140a-49a2-bde5-f220b05aa252" (UID: "05c9a6ea-140a-49a2-bde5-f220b05aa252"). InnerVolumeSpecName "secret-alertmanager-kube-rbac-proxy-metric". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.156473 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-main-proxy" (OuterVolumeSpecName: "secret-alertmanager-main-proxy") pod "05c9a6ea-140a-49a2-bde5-f220b05aa252" (UID: "05c9a6ea-140a-49a2-bde5-f220b05aa252"). InnerVolumeSpecName "secret-alertmanager-main-proxy". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.159906 2588 reconciler.go:399] "Volume detached for volume \"secret-alertmanager-main-proxy\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-main-proxy\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.159938 2588 reconciler.go:399] "Volume detached for volume \"kube-api-access-kxsm5\" (UniqueName: \"kubernetes.io/projected/05c9a6ea-140a-49a2-bde5-f220b05aa252-kube-api-access-kxsm5\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.159956 2588 reconciler.go:399] "Volume detached for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/05c9a6ea-140a-49a2-bde5-f220b05aa252-metrics-client-ca\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.159974 2588 reconciler.go:399] "Volume detached for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/05c9a6ea-140a-49a2-bde5-f220b05aa252-alertmanager-main-db\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.159997 2588 reconciler.go:399] "Volume detached for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-kube-rbac-proxy-metric\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.160014 2588 reconciler.go:399] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/05c9a6ea-140a-49a2-bde5-f220b05aa252-tls-assets\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.160033 2588 reconciler.go:399] "Volume detached for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05c9a6ea-140a-49a2-bde5-f220b05aa252-alertmanager-trusted-ca-bundle\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.160050 2588 reconciler.go:399] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-config-volume\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.160066 2588 reconciler.go:399] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/05c9a6ea-140a-49a2-bde5-f220b05aa252-config-out\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.160084 2588 reconciler.go:399] "Volume detached for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-kube-rbac-proxy\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.180500 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-web-config" (OuterVolumeSpecName: "web-config") pod "05c9a6ea-140a-49a2-bde5-f220b05aa252" (UID: "05c9a6ea-140a-49a2-bde5-f220b05aa252"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.200617 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-main-tls" (OuterVolumeSpecName: "secret-alertmanager-main-tls") pod "05c9a6ea-140a-49a2-bde5-f220b05aa252" (UID: "05c9a6ea-140a-49a2-bde5-f220b05aa252"). InnerVolumeSpecName "secret-alertmanager-main-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.205730077Z" level=info msg="Removed container 3cbc4c089943552285b7235ee6268116978902d078a993e4f88285e0c16777ce: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy-rules" id=dd679a59-d8e8-4b2a-aba6-826719daa3ee name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.205930 2588 scope.go:115] "RemoveContainer" containerID="b7574d11ea7895beed8a146040a8adb91835493e426424dd8e461a93b6f7677b" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.209732415Z" level=info msg="Removing container: b7574d11ea7895beed8a146040a8adb91835493e426424dd8e461a93b6f7677b" id=66e25fca-6944-4a42-8fa2-9da2ea0a30c1 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.233631078Z" level=info msg="Removed container b7574d11ea7895beed8a146040a8adb91835493e426424dd8e461a93b6f7677b: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/prom-label-proxy" id=66e25fca-6944-4a42-8fa2-9da2ea0a30c1 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.233792 2588 scope.go:115] "RemoveContainer" containerID="c507c8393ab5e7e8bb6dbc97bcabc3f2598889f9b13fbee87614f2fee17b1e9d" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.234583438Z" level=info msg="Removing container: c507c8393ab5e7e8bb6dbc97bcabc3f2598889f9b13fbee87614f2fee17b1e9d" id=e469bf8e-709b-41ca-b4f0-f42e5e0c0b86 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.257937082Z" level=info msg="Removed container c507c8393ab5e7e8bb6dbc97bcabc3f2598889f9b13fbee87614f2fee17b1e9d: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/kube-rbac-proxy" id=e469bf8e-709b-41ca-b4f0-f42e5e0c0b86 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.258088 2588 scope.go:115] "RemoveContainer" containerID="76d0bbfe3b04b3fb31c951bf14c59dddccfd473352dd9388eeee0e7e51e96358" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.258845399Z" level=info msg="Removing container: 76d0bbfe3b04b3fb31c951bf14c59dddccfd473352dd9388eeee0e7e51e96358" id=2cbe973c-6733-44b2-8808-8b1c30208a30 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.260740 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-telemeter-client\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-secret-telemeter-client\") pod \"fc45e847-8f55-4c48-8c48-4bfee2419dca\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.260790 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4r64p\" (UniqueName: \"kubernetes.io/projected/fc45e847-8f55-4c48-8c48-4bfee2419dca-kube-api-access-4r64p\") pod \"fc45e847-8f55-4c48-8c48-4bfee2419dca\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.260824 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"telemeter-client-tls\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-telemeter-client-tls\") pod \"fc45e847-8f55-4c48-8c48-4bfee2419dca\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.260854 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-metrics-client-ca\") pod \"fc45e847-8f55-4c48-8c48-4bfee2419dca\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.260890 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"telemeter-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-telemeter-trusted-ca-bundle\") pod \"fc45e847-8f55-4c48-8c48-4bfee2419dca\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.260924 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-serving-certs-ca-bundle\") pod \"fc45e847-8f55-4c48-8c48-4bfee2419dca\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.260958 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-telemeter-client-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-secret-telemeter-client-kube-rbac-proxy-config\") pod \"fc45e847-8f55-4c48-8c48-4bfee2419dca\" (UID: \"fc45e847-8f55-4c48-8c48-4bfee2419dca\") " Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.261086 2588 reconciler.go:399] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-web-config\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.261110 2588 reconciler.go:399] "Volume detached for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/05c9a6ea-140a-49a2-bde5-f220b05aa252-secret-alertmanager-main-tls\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:44:50.263017 2588 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/fc45e847-8f55-4c48-8c48-4bfee2419dca/volumes/kubernetes.io~configmap/telemeter-trusted-ca-bundle: clearQuota called, but quotas disabled Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.263281 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-telemeter-trusted-ca-bundle" (OuterVolumeSpecName: "telemeter-trusted-ca-bundle") pod "fc45e847-8f55-4c48-8c48-4bfee2419dca" (UID: "fc45e847-8f55-4c48-8c48-4bfee2419dca"). InnerVolumeSpecName "telemeter-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:44:50.263435 2588 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/fc45e847-8f55-4c48-8c48-4bfee2419dca/volumes/kubernetes.io~configmap/metrics-client-ca: clearQuota called, but quotas disabled Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.263617 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-metrics-client-ca" (OuterVolumeSpecName: "metrics-client-ca") pod "fc45e847-8f55-4c48-8c48-4bfee2419dca" (UID: "fc45e847-8f55-4c48-8c48-4bfee2419dca"). InnerVolumeSpecName "metrics-client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:44:50.263750 2588 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/fc45e847-8f55-4c48-8c48-4bfee2419dca/volumes/kubernetes.io~configmap/serving-certs-ca-bundle: clearQuota called, but quotas disabled Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.263916 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-serving-certs-ca-bundle" (OuterVolumeSpecName: "serving-certs-ca-bundle") pod "fc45e847-8f55-4c48-8c48-4bfee2419dca" (UID: "fc45e847-8f55-4c48-8c48-4bfee2419dca"). InnerVolumeSpecName "serving-certs-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.276824 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-secret-telemeter-client-kube-rbac-proxy-config" (OuterVolumeSpecName: "secret-telemeter-client-kube-rbac-proxy-config") pod "fc45e847-8f55-4c48-8c48-4bfee2419dca" (UID: "fc45e847-8f55-4c48-8c48-4bfee2419dca"). InnerVolumeSpecName "secret-telemeter-client-kube-rbac-proxy-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.284510 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc45e847-8f55-4c48-8c48-4bfee2419dca-kube-api-access-4r64p" (OuterVolumeSpecName: "kube-api-access-4r64p") pod "fc45e847-8f55-4c48-8c48-4bfee2419dca" (UID: "fc45e847-8f55-4c48-8c48-4bfee2419dca"). InnerVolumeSpecName "kube-api-access-4r64p". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.285085014Z" level=info msg="Removed container 76d0bbfe3b04b3fb31c951bf14c59dddccfd473352dd9388eeee0e7e51e96358: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/oauth-proxy" id=2cbe973c-6733-44b2-8808-8b1c30208a30 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.285538 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-telemeter-client-tls" (OuterVolumeSpecName: "telemeter-client-tls") pod "fc45e847-8f55-4c48-8c48-4bfee2419dca" (UID: "fc45e847-8f55-4c48-8c48-4bfee2419dca"). InnerVolumeSpecName "telemeter-client-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.286461 2588 scope.go:115] "RemoveContainer" containerID="578cce0ed1aa4795c3b6d8648ecf4b937233d871e6a027eb233bec20d4333c1f" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.287204514Z" level=info msg="Removing container: 578cce0ed1aa4795c3b6d8648ecf4b937233d871e6a027eb233bec20d4333c1f" id=8446040e-6697-4d74-963b-e211487bdd16 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.290520 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-secret-telemeter-client" (OuterVolumeSpecName: "secret-telemeter-client") pod "fc45e847-8f55-4c48-8c48-4bfee2419dca" (UID: "fc45e847-8f55-4c48-8c48-4bfee2419dca"). InnerVolumeSpecName "secret-telemeter-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.304426113Z" level=info msg="Removed container 578cce0ed1aa4795c3b6d8648ecf4b937233d871e6a027eb233bec20d4333c1f: openshift-monitoring/thanos-querier-6f4f5fb7cf-rdm9l/thanos-query" id=8446040e-6697-4d74-963b-e211487bdd16 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.304607 2588 scope.go:115] "RemoveContainer" containerID="7eca8a735d65a0f0812794e42a36fd300146f8782a35f3df7830a91dfdd49bef" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:44:50.304885 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7eca8a735d65a0f0812794e42a36fd300146f8782a35f3df7830a91dfdd49bef\": container with ID starting with 7eca8a735d65a0f0812794e42a36fd300146f8782a35f3df7830a91dfdd49bef not found: ID does not exist" containerID="7eca8a735d65a0f0812794e42a36fd300146f8782a35f3df7830a91dfdd49bef" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.304921 2588 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:7eca8a735d65a0f0812794e42a36fd300146f8782a35f3df7830a91dfdd49bef} err="failed to get container status \"7eca8a735d65a0f0812794e42a36fd300146f8782a35f3df7830a91dfdd49bef\": rpc error: code = NotFound desc = could not find container \"7eca8a735d65a0f0812794e42a36fd300146f8782a35f3df7830a91dfdd49bef\": container with ID starting with 7eca8a735d65a0f0812794e42a36fd300146f8782a35f3df7830a91dfdd49bef not found: ID does not exist" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.304936 2588 scope.go:115] "RemoveContainer" containerID="3cbc4c089943552285b7235ee6268116978902d078a993e4f88285e0c16777ce" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:44:50.305256 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cbc4c089943552285b7235ee6268116978902d078a993e4f88285e0c16777ce\": container with ID starting with 3cbc4c089943552285b7235ee6268116978902d078a993e4f88285e0c16777ce not found: ID does not exist" containerID="3cbc4c089943552285b7235ee6268116978902d078a993e4f88285e0c16777ce" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.305300 2588 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:3cbc4c089943552285b7235ee6268116978902d078a993e4f88285e0c16777ce} err="failed to get container status \"3cbc4c089943552285b7235ee6268116978902d078a993e4f88285e0c16777ce\": rpc error: code = NotFound desc = could not find container \"3cbc4c089943552285b7235ee6268116978902d078a993e4f88285e0c16777ce\": container with ID starting with 3cbc4c089943552285b7235ee6268116978902d078a993e4f88285e0c16777ce not found: ID does not exist" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.305309 2588 scope.go:115] "RemoveContainer" containerID="b7574d11ea7895beed8a146040a8adb91835493e426424dd8e461a93b6f7677b" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:44:50.305549 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7574d11ea7895beed8a146040a8adb91835493e426424dd8e461a93b6f7677b\": container with ID starting with b7574d11ea7895beed8a146040a8adb91835493e426424dd8e461a93b6f7677b not found: ID does not exist" containerID="b7574d11ea7895beed8a146040a8adb91835493e426424dd8e461a93b6f7677b" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.305579 2588 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:b7574d11ea7895beed8a146040a8adb91835493e426424dd8e461a93b6f7677b} err="failed to get container status \"b7574d11ea7895beed8a146040a8adb91835493e426424dd8e461a93b6f7677b\": rpc error: code = NotFound desc = could not find container \"b7574d11ea7895beed8a146040a8adb91835493e426424dd8e461a93b6f7677b\": container with ID starting with b7574d11ea7895beed8a146040a8adb91835493e426424dd8e461a93b6f7677b not found: ID does not exist" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.305593 2588 scope.go:115] "RemoveContainer" containerID="c507c8393ab5e7e8bb6dbc97bcabc3f2598889f9b13fbee87614f2fee17b1e9d" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:44:50.305792 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c507c8393ab5e7e8bb6dbc97bcabc3f2598889f9b13fbee87614f2fee17b1e9d\": container with ID starting with c507c8393ab5e7e8bb6dbc97bcabc3f2598889f9b13fbee87614f2fee17b1e9d not found: ID does not exist" containerID="c507c8393ab5e7e8bb6dbc97bcabc3f2598889f9b13fbee87614f2fee17b1e9d" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.305822 2588 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:c507c8393ab5e7e8bb6dbc97bcabc3f2598889f9b13fbee87614f2fee17b1e9d} err="failed to get container status \"c507c8393ab5e7e8bb6dbc97bcabc3f2598889f9b13fbee87614f2fee17b1e9d\": rpc error: code = NotFound desc = could not find container \"c507c8393ab5e7e8bb6dbc97bcabc3f2598889f9b13fbee87614f2fee17b1e9d\": container with ID starting with c507c8393ab5e7e8bb6dbc97bcabc3f2598889f9b13fbee87614f2fee17b1e9d not found: ID does not exist" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.305835 2588 scope.go:115] "RemoveContainer" containerID="76d0bbfe3b04b3fb31c951bf14c59dddccfd473352dd9388eeee0e7e51e96358" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:44:50.306033 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76d0bbfe3b04b3fb31c951bf14c59dddccfd473352dd9388eeee0e7e51e96358\": container with ID starting with 76d0bbfe3b04b3fb31c951bf14c59dddccfd473352dd9388eeee0e7e51e96358 not found: ID does not exist" containerID="76d0bbfe3b04b3fb31c951bf14c59dddccfd473352dd9388eeee0e7e51e96358" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.306067 2588 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:76d0bbfe3b04b3fb31c951bf14c59dddccfd473352dd9388eeee0e7e51e96358} err="failed to get container status \"76d0bbfe3b04b3fb31c951bf14c59dddccfd473352dd9388eeee0e7e51e96358\": rpc error: code = NotFound desc = could not find container \"76d0bbfe3b04b3fb31c951bf14c59dddccfd473352dd9388eeee0e7e51e96358\": container with ID starting with 76d0bbfe3b04b3fb31c951bf14c59dddccfd473352dd9388eeee0e7e51e96358 not found: ID does not exist" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.306081 2588 scope.go:115] "RemoveContainer" containerID="578cce0ed1aa4795c3b6d8648ecf4b937233d871e6a027eb233bec20d4333c1f" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:44:50.306341 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"578cce0ed1aa4795c3b6d8648ecf4b937233d871e6a027eb233bec20d4333c1f\": container with ID starting with 578cce0ed1aa4795c3b6d8648ecf4b937233d871e6a027eb233bec20d4333c1f not found: ID does not exist" containerID="578cce0ed1aa4795c3b6d8648ecf4b937233d871e6a027eb233bec20d4333c1f" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.306370 2588 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:578cce0ed1aa4795c3b6d8648ecf4b937233d871e6a027eb233bec20d4333c1f} err="failed to get container status \"578cce0ed1aa4795c3b6d8648ecf4b937233d871e6a027eb233bec20d4333c1f\": rpc error: code = NotFound desc = could not find container \"578cce0ed1aa4795c3b6d8648ecf4b937233d871e6a027eb233bec20d4333c1f\": container with ID starting with 578cce0ed1aa4795c3b6d8648ecf4b937233d871e6a027eb233bec20d4333c1f not found: ID does not exist" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.352977 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh" event=&{ID:a40bf20d-67a4-45e7-b701-ca5a75376c1f Type:ContainerDied Data:9d9807a6998df2507a4aa5b9bad79e347bc60d62b254989738c927929d827e6e} Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.353008 2588 scope.go:115] "RemoveContainer" containerID="543f2356973a019ec622df9cf319f02abc2eac1228b813bef15ed997a118fe6d" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.355121522Z" level=info msg="Removing container: 543f2356973a019ec622df9cf319f02abc2eac1228b813bef15ed997a118fe6d" id=78196035-aa86-4392-8c29-55613dba565a name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.356775 2588 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_05c9a6ea-140a-49a2-bde5-f220b05aa252/alertmanager-proxy/2.log" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.357338 2588 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_05c9a6ea-140a-49a2-bde5-f220b05aa252/config-reloader/2.log" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.358581 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event=&{ID:05c9a6ea-140a-49a2-bde5-f220b05aa252 Type:ContainerDied Data:d6d0c82bc9774e51559f70121d26a57c7c38cce6fb83f23b63a8bb671c408015} Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.363741 2588 reconciler.go:399] "Volume detached for volume \"secret-telemeter-client\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-secret-telemeter-client\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.363767 2588 reconciler.go:399] "Volume detached for volume \"kube-api-access-4r64p\" (UniqueName: \"kubernetes.io/projected/fc45e847-8f55-4c48-8c48-4bfee2419dca-kube-api-access-4r64p\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.363786 2588 reconciler.go:399] "Volume detached for volume \"telemeter-client-tls\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-telemeter-client-tls\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.363803 2588 reconciler.go:399] "Volume detached for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-metrics-client-ca\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.363821 2588 reconciler.go:399] "Volume detached for volume \"telemeter-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-telemeter-trusted-ca-bundle\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.363862 2588 reconciler.go:399] "Volume detached for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fc45e847-8f55-4c48-8c48-4bfee2419dca-serving-certs-ca-bundle\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.363885 2588 reconciler.go:399] "Volume detached for volume \"secret-telemeter-client-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/fc45e847-8f55-4c48-8c48-4bfee2419dca-secret-telemeter-client-kube-rbac-proxy-config\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.374495 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-5598c645c4-mftdv" event=&{ID:fc45e847-8f55-4c48-8c48-4bfee2419dca Type:ContainerDied Data:f9cc29d007bd3c948c46f781abd08afdb46359ec50fc31f8c503fcdb35b7aa22} Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.439851740Z" level=info msg="Removed container 543f2356973a019ec622df9cf319f02abc2eac1228b813bef15ed997a118fe6d: openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh/prometheus-adapter" id=78196035-aa86-4392-8c29-55613dba565a name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.440209 2588 scope.go:115] "RemoveContainer" containerID="4a95d41273862399b3cca16471b999defcccef3ceb8402d7cfb798378ffc8cab" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.441339269Z" level=info msg="Removing container: 4a95d41273862399b3cca16471b999defcccef3ceb8402d7cfb798378ffc8cab" id=3ada6d80-aacf-42a1-8bb0-1356833cad88 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.477719 2588 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-monitoring/telemeter-client-5598c645c4-mftdv] Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.482637 2588 kubelet.go:2129] "SyncLoop REMOVE" source="api" pods=[openshift-monitoring/telemeter-client-5598c645c4-mftdv] Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.483109184Z" level=info msg="Removed container 4a95d41273862399b3cca16471b999defcccef3ceb8402d7cfb798378ffc8cab: openshift-monitoring/alertmanager-main-0/prom-label-proxy" id=3ada6d80-aacf-42a1-8bb0-1356833cad88 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.484688 2588 scope.go:115] "RemoveContainer" containerID="4c970eddded2809e2e506cfb0b207a19b375bc49b3a1f9bfd58b202b7c1f6ebd" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.485729580Z" level=info msg="Removing container: 4c970eddded2809e2e506cfb0b207a19b375bc49b3a1f9bfd58b202b7c1f6ebd" id=11a43076-2b79-4e7b-bc37-d41ac241625a name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.506644 2588 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh] Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.510717 2588 kubelet.go:2129] "SyncLoop REMOVE" source="api" pods=[openshift-monitoring/prometheus-adapter-5894cc86c7-qhmbh] Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.526326015Z" level=info msg="Removed container 4c970eddded2809e2e506cfb0b207a19b375bc49b3a1f9bfd58b202b7c1f6ebd: openshift-monitoring/alertmanager-main-0/kube-rbac-proxy-metric" id=11a43076-2b79-4e7b-bc37-d41ac241625a name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.526526 2588 scope.go:115] "RemoveContainer" containerID="53a26d2c619d3ef8230bda385bc9245dc43ada80f94fe54f1f79b82763d2f7a8" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.527292616Z" level=info msg="Removing container: 53a26d2c619d3ef8230bda385bc9245dc43ada80f94fe54f1f79b82763d2f7a8" id=0f8f89fd-1145-4632-b844-005dc3e05dfc name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.532810 2588 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-monitoring/alertmanager-main-0] Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.539800 2588 kubelet.go:2129] "SyncLoop REMOVE" source="api" pods=[openshift-monitoring/alertmanager-main-0] Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.570798999Z" level=info msg="Removed container 53a26d2c619d3ef8230bda385bc9245dc43ada80f94fe54f1f79b82763d2f7a8: openshift-monitoring/alertmanager-main-0/kube-rbac-proxy" id=0f8f89fd-1145-4632-b844-005dc3e05dfc name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.571121 2588 scope.go:115] "RemoveContainer" containerID="85c843961a0e8a2283e240b3c350b3d3823e13fbe3cb8f291a0e809df84275e7" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.572251304Z" level=info msg="Removing container: 85c843961a0e8a2283e240b3c350b3d3823e13fbe3cb8f291a0e809df84275e7" id=b6c045b5-9891-4ed2-ac2e-3d0405ccd0a8 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.582182 2588 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeNotSchedulable" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.605514896Z" level=info msg="Removed container 85c843961a0e8a2283e240b3c350b3d3823e13fbe3cb8f291a0e809df84275e7: openshift-monitoring/alertmanager-main-0/alertmanager-proxy" id=b6c045b5-9891-4ed2-ac2e-3d0405ccd0a8 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.605693 2588 scope.go:115] "RemoveContainer" containerID="df204cb60308dd066ca32f1f38b7a71d29064341d03539fedabfb840583c8a96" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.608217867Z" level=info msg="Removing container: df204cb60308dd066ca32f1f38b7a71d29064341d03539fedabfb840583c8a96" id=f2d41671-e677-42d0-a0e0-7ae13f81fca7 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.633571820Z" level=info msg="Removed container df204cb60308dd066ca32f1f38b7a71d29064341d03539fedabfb840583c8a96: openshift-monitoring/alertmanager-main-0/config-reloader" id=f2d41671-e677-42d0-a0e0-7ae13f81fca7 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.634056 2588 scope.go:115] "RemoveContainer" containerID="29d242ba69edfb0ca301e890f9169a7355a0889dbda9e344277448360467524b" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.635195981Z" level=info msg="Removing container: 29d242ba69edfb0ca301e890f9169a7355a0889dbda9e344277448360467524b" id=8afbbcf9-06ec-474f-b084-a0b2a04f9007 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.662814907Z" level=info msg="Removed container 29d242ba69edfb0ca301e890f9169a7355a0889dbda9e344277448360467524b: openshift-monitoring/alertmanager-main-0/alertmanager" id=8afbbcf9-06ec-474f-b084-a0b2a04f9007 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.663462 2588 scope.go:115] "RemoveContainer" containerID="91947bee1cff242d8ff480a4ba98a9c65dd41326ef2bfc74ccc1a969906e3b00" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.664561136Z" level=info msg="Removing container: 91947bee1cff242d8ff480a4ba98a9c65dd41326ef2bfc74ccc1a969906e3b00" id=4fe51824-e0c0-4861-9fcd-1f16f90d01ce name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.693965565Z" level=info msg="Removed container 91947bee1cff242d8ff480a4ba98a9c65dd41326ef2bfc74ccc1a969906e3b00: openshift-monitoring/telemeter-client-5598c645c4-mftdv/kube-rbac-proxy" id=4fe51824-e0c0-4861-9fcd-1f16f90d01ce name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.694483 2588 scope.go:115] "RemoveContainer" containerID="2952a168f0605321dd4ff798bf4481d0ede73526e359849e9e7039226b853fab" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.697151644Z" level=info msg="Removing container: 2952a168f0605321dd4ff798bf4481d0ede73526e359849e9e7039226b853fab" id=cafb369a-6dec-4ee6-a6de-a84a4aa3df35 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.725401592Z" level=info msg="Removed container 2952a168f0605321dd4ff798bf4481d0ede73526e359849e9e7039226b853fab: openshift-monitoring/telemeter-client-5598c645c4-mftdv/reload" id=cafb369a-6dec-4ee6-a6de-a84a4aa3df35 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.726522920Z" level=info msg="Removing container: 2c1936a19655345a892373f86a623f7409e46829cbb1adeca75e2879d268aa08" id=ba2595cf-67f9-4946-9a53-a7788f29bfeb name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.725590 2588 scope.go:115] "RemoveContainer" containerID="2c1936a19655345a892373f86a623f7409e46829cbb1adeca75e2879d268aa08" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.800258 2588 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=05c9a6ea-140a-49a2-bde5-f220b05aa252 path="/var/lib/kubelet/pods/05c9a6ea-140a-49a2-bde5-f220b05aa252/volumes" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.801584 2588 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=08c04948-e189-4227-bece-151bf3576721 path="/var/lib/kubelet/pods/08c04948-e189-4227-bece-151bf3576721/volumes" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.802246 2588 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=2ecd2e2c-ab52-42e0-8fde-64e45804cc9e path="/var/lib/kubelet/pods/2ecd2e2c-ab52-42e0-8fde-64e45804cc9e/volumes" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.803186 2588 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=7c63c9be-d1ad-44b6-8b33-09af3ee314af path="/var/lib/kubelet/pods/7c63c9be-d1ad-44b6-8b33-09af3ee314af/volumes" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.808256 2588 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=88745ab7-efe7-42c0-a2e8-75591518333d path="/var/lib/kubelet/pods/88745ab7-efe7-42c0-a2e8-75591518333d/volumes" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.808741 2588 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=9de5de36-9b80-4a0b-a615-f0f9ccda2ff6 path="/var/lib/kubelet/pods/9de5de36-9b80-4a0b-a615-f0f9ccda2ff6/volumes" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.809439 2588 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=a40bf20d-67a4-45e7-b701-ca5a75376c1f path="/var/lib/kubelet/pods/a40bf20d-67a4-45e7-b701-ca5a75376c1f/volumes" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:50.809931 2588 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=fc45e847-8f55-4c48-8c48-4bfee2419dca path="/var/lib/kubelet/pods/fc45e847-8f55-4c48-8c48-4bfee2419dca/volumes" Jan 05 09:44:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:44:50.811385192Z" level=info msg="Removed container 2c1936a19655345a892373f86a623f7409e46829cbb1adeca75e2879d268aa08: openshift-monitoring/telemeter-client-5598c645c4-mftdv/telemeter-client" id=ba2595cf-67f9-4946-9a53-a7788f29bfeb name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:44:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:44:52.346563 2588 remote_runtime.go:734] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73 is running failed: container process not found" containerID="4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73" cmd=[sh -c if [ -x "$(command -v curl)" ]; then exec curl --fail http://localhost:9090/-/ready; elif [ -x "$(command -v wget)" ]; then exec wget -q -O /dev/null http://localhost:9090/-/ready; else exit 1; fi] Jan 05 09:44:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:44:52.347203 2588 remote_runtime.go:734] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73 is running failed: container process not found" containerID="4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73" cmd=[sh -c if [ -x "$(command -v curl)" ]; then exec curl --fail http://localhost:9090/-/ready; elif [ -x "$(command -v wget)" ]; then exec wget -q -O /dev/null http://localhost:9090/-/ready; else exit 1; fi] Jan 05 09:44:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:44:52.347459 2588 remote_runtime.go:734] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73 is running failed: container process not found" containerID="4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73" cmd=[sh -c if [ -x "$(command -v curl)" ]; then exec curl --fail http://localhost:9090/-/ready; elif [ -x "$(command -v wget)" ]; then exec wget -q -O /dev/null http://localhost:9090/-/ready; else exit 1; fi] Jan 05 09:44:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:44:52.347504 2588 prober.go:111] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73 is running failed: container process not found" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID=4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 containerName="prometheus" Jan 05 09:44:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:52.444550 2588 patch_prober.go:29] interesting pod/router-default-5cf6bbdff9-tnffs container/router namespace/openshift-ingress: Readiness probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]backend-http ok Jan 05 09:44:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: [+]has-synced ok Jan 05 09:44:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: [-]process-running failed: reason withheld Jan 05 09:44:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: healthz check failed Jan 05 09:44:52 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:44:52.444603 2588 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" podUID=c3b10ff4-a243-4a4e-b27d-f613e258666d containerName="router" probeResult=failure output="HTTP probe failed with statuscode: 500" Jan 05 09:44:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:44:57.347631 2588 remote_runtime.go:734] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73 is running failed: container process not found" containerID="4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73" cmd=[sh -c if [ -x "$(command -v curl)" ]; then exec curl --fail http://localhost:9090/-/ready; elif [ -x "$(command -v wget)" ]; then exec wget -q -O /dev/null http://localhost:9090/-/ready; else exit 1; fi] Jan 05 09:44:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:44:57.349884 2588 remote_runtime.go:734] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73 is running failed: container process not found" containerID="4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73" cmd=[sh -c if [ -x "$(command -v curl)" ]; then exec curl --fail http://localhost:9090/-/ready; elif [ -x "$(command -v wget)" ]; then exec wget -q -O /dev/null http://localhost:9090/-/ready; else exit 1; fi] Jan 05 09:44:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:44:57.350391 2588 remote_runtime.go:734] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73 is running failed: container process not found" containerID="4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73" cmd=[sh -c if [ -x "$(command -v curl)" ]; then exec curl --fail http://localhost:9090/-/ready; elif [ -x "$(command -v wget)" ]; then exec wget -q -O /dev/null http://localhost:9090/-/ready; else exit 1; fi] Jan 05 09:44:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:44:57.350432 2588 prober.go:111] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73 is running failed: container process not found" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID=4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 containerName="prometheus" Jan 05 09:45:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:02.346872 2588 remote_runtime.go:734] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73 is running failed: container process not found" containerID="4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73" cmd=[sh -c if [ -x "$(command -v curl)" ]; then exec curl --fail http://localhost:9090/-/ready; elif [ -x "$(command -v wget)" ]; then exec wget -q -O /dev/null http://localhost:9090/-/ready; else exit 1; fi] Jan 05 09:45:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:02.347213 2588 remote_runtime.go:734] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73 is running failed: container process not found" containerID="4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73" cmd=[sh -c if [ -x "$(command -v curl)" ]; then exec curl --fail http://localhost:9090/-/ready; elif [ -x "$(command -v wget)" ]; then exec wget -q -O /dev/null http://localhost:9090/-/ready; else exit 1; fi] Jan 05 09:45:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:02.347482 2588 remote_runtime.go:734] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73 is running failed: container process not found" containerID="4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73" cmd=[sh -c if [ -x "$(command -v curl)" ]; then exec curl --fail http://localhost:9090/-/ready; elif [ -x "$(command -v wget)" ]; then exec wget -q -O /dev/null http://localhost:9090/-/ready; else exit 1; fi] Jan 05 09:45:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:02.347521 2588 prober.go:111] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73 is running failed: container process not found" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID=4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 containerName="prometheus" Jan 05 09:45:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:02.444612 2588 patch_prober.go:29] interesting pod/router-default-5cf6bbdff9-tnffs container/router namespace/openshift-ingress: Readiness probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]backend-http ok Jan 05 09:45:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: [+]has-synced ok Jan 05 09:45:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: [-]process-running failed: reason withheld Jan 05 09:45:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: healthz check failed Jan 05 09:45:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:02.444683 2588 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" podUID=c3b10ff4-a243-4a4e-b27d-f613e258666d containerName="router" probeResult=failure output="HTTP probe failed with statuscode: 500" Jan 05 09:45:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:04.726790823Z" level=error msg="Failed to cleanup (probably retrying): failed to destroy network for pod sandbox k8s_prometheus-k8s-0_openshift-monitoring_4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879_0(04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0): error removing pod openshift-monitoring_prometheus-k8s-0 from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-monitoring/prometheus-k8s-0/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" Jan 05 09:45:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:04.726894875Z" level=info msg="Got pod network &{Name:prometheus-k8s-0 Namespace:openshift-monitoring ID:575fca5742eeb550505988169b65e39889d4e9a4a916a669f68c988db3eb4df1 UID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 NetNS:/var/run/netns/1cd4faa3-1dec-4efe-a53a-6ae0a54cc872 Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:45:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:04.727060547Z" level=info msg="Deleting pod openshift-monitoring_prometheus-k8s-0 from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:45:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:04.823645 2588 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/a291ecea7e611cbd312be7038a1070e9c00b70f193e0861dfa9475efb74b68e8/diff" to get inode usage: stat /var/lib/containers/storage/overlay/a291ecea7e611cbd312be7038a1070e9c00b70f193e0861dfa9475efb74b68e8/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openshift-kube-storage-version-migrator_migrator-6cd87bdff8-wj8dg_08c04948-e189-4227-bece-151bf3576721/migrator/0.log" to get inode usage: stat /var/log/pods/openshift-kube-storage-version-migrator_migrator-6cd87bdff8-wj8dg_08c04948-e189-4227-bece-151bf3576721/migrator/0.log: no such file or directory Jan 05 09:45:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:45:04Z [verbose] Del: openshift-monitoring:prometheus-k8s-0:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:45:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:04.898659012Z" level=info msg="Stopped pod sandbox: 575fca5742eeb550505988169b65e39889d4e9a4a916a669f68c988db3eb4df1" id=5c36e6c8-5cf7-4e53-9089-5c5cc2eefa37 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:04.907447 2588 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879/prometheus-proxy/2.log" Jan 05 09:45:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:04.908048 2588 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879/config-reloader/2.log" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.070298 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-prometheus-k8s-proxy\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-proxy\") pod \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.070350 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-trusted-ca-bundle\") pod \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.070382 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.070411 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-config-out\") pod \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.070437 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-tls-assets\") pod \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.070469 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-kube-etcd-client-certs\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-kube-etcd-client-certs\") pod \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.070494 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-config\") pod \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.070523 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-web-config\") pod \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.070557 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-kubelet-serving-ca-bundle\") pod \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.070583 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-metrics-client-ca\") pod \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.070617 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-metrics-client-certs\") pod \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.070644 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98gp9\" (UniqueName: \"kubernetes.io/projected/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-kube-api-access-98gp9\") pod \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.070679 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-kube-rbac-proxy\") pod \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.070708 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-grpc-tls\") pod \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.070740 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-metrics-client-ca\") pod \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.070778 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-tls\") pod \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.070804 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-k8s-rulefiles-0\") pod \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.070837 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-serving-certs-ca-bundle\") pod \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.070884 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-k8s-db\") pod \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\" (UID: \"4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879\") " Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:45:05.071208 2588 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879/volumes/kubernetes.io~empty-dir/prometheus-k8s-db: clearQuota called, but quotas disabled Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.075953 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-k8s-db" (OuterVolumeSpecName: "prometheus-k8s-db") pod "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879" (UID: "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879"). InnerVolumeSpecName "prometheus-k8s-db". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:45:05.076181 2588 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879/volumes/kubernetes.io~configmap/prometheus-trusted-ca-bundle: clearQuota called, but quotas disabled Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.076408 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-trusted-ca-bundle" (OuterVolumeSpecName: "prometheus-trusted-ca-bundle") pod "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879" (UID: "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879"). InnerVolumeSpecName "prometheus-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:45:05.088741 2588 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879/volumes/kubernetes.io~empty-dir/config-out: clearQuota called, but quotas disabled Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.088877 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-config-out" (OuterVolumeSpecName: "config-out") pod "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879" (UID: "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:45:05.089662 2588 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879/volumes/kubernetes.io~configmap/configmap-metrics-client-ca: clearQuota called, but quotas disabled Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.089854 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-metrics-client-ca" (OuterVolumeSpecName: "configmap-metrics-client-ca") pod "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879" (UID: "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879"). InnerVolumeSpecName "configmap-metrics-client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:45:05.090241 2588 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879/volumes/kubernetes.io~configmap/prometheus-k8s-rulefiles-0: clearQuota called, but quotas disabled Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.092475 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-proxy" (OuterVolumeSpecName: "secret-prometheus-k8s-proxy") pod "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879" (UID: "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879"). InnerVolumeSpecName "secret-prometheus-k8s-proxy". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.092504 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-k8s-rulefiles-0" (OuterVolumeSpecName: "prometheus-k8s-rulefiles-0") pod "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879" (UID: "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879"). InnerVolumeSpecName "prometheus-k8s-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:45:05.092651 2588 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879/volumes/kubernetes.io~configmap/configmap-serving-certs-ca-bundle: clearQuota called, but quotas disabled Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.092816 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-serving-certs-ca-bundle" (OuterVolumeSpecName: "configmap-serving-certs-ca-bundle") pod "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879" (UID: "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879"). InnerVolumeSpecName "configmap-serving-certs-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:45:05.092966 2588 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879/volumes/kubernetes.io~configmap/configmap-kubelet-serving-ca-bundle: clearQuota called, but quotas disabled Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.093178 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-kubelet-serving-ca-bundle" (OuterVolumeSpecName: "configmap-kubelet-serving-ca-bundle") pod "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879" (UID: "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879"). InnerVolumeSpecName "configmap-kubelet-serving-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:45:05.093682 2588 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879/volumes/kubernetes.io~configmap/metrics-client-ca: clearQuota called, but quotas disabled Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.093890 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-metrics-client-ca" (OuterVolumeSpecName: "metrics-client-ca") pod "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879" (UID: "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879"). InnerVolumeSpecName "metrics-client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.098247 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-thanos-sidecar-tls" (OuterVolumeSpecName: "secret-prometheus-k8s-thanos-sidecar-tls") pod "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879" (UID: "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879"). InnerVolumeSpecName "secret-prometheus-k8s-thanos-sidecar-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.101817 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-metrics-client-certs" (OuterVolumeSpecName: "secret-metrics-client-certs") pod "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879" (UID: "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879"). InnerVolumeSpecName "secret-metrics-client-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.101979 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879" (UID: "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.104458 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-tls" (OuterVolumeSpecName: "secret-prometheus-k8s-tls") pod "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879" (UID: "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879"). InnerVolumeSpecName "secret-prometheus-k8s-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.104659 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-kube-api-access-98gp9" (OuterVolumeSpecName: "kube-api-access-98gp9") pod "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879" (UID: "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879"). InnerVolumeSpecName "kube-api-access-98gp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.106452 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-config" (OuterVolumeSpecName: "config") pod "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879" (UID: "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.110421 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-kube-etcd-client-certs" (OuterVolumeSpecName: "secret-kube-etcd-client-certs") pod "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879" (UID: "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879"). InnerVolumeSpecName "secret-kube-etcd-client-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.114424 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-kube-rbac-proxy" (OuterVolumeSpecName: "secret-kube-rbac-proxy") pod "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879" (UID: "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879"). InnerVolumeSpecName "secret-kube-rbac-proxy". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.117414 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-grpc-tls" (OuterVolumeSpecName: "secret-grpc-tls") pod "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879" (UID: "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879"). InnerVolumeSpecName "secret-grpc-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.121392 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-web-config" (OuterVolumeSpecName: "web-config") pod "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879" (UID: "4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.171896 2588 reconciler.go:399] "Volume detached for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-serving-certs-ca-bundle\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.171921 2588 reconciler.go:399] "Volume detached for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-k8s-db\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.171940 2588 reconciler.go:399] "Volume detached for volume \"secret-prometheus-k8s-proxy\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-proxy\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.171957 2588 reconciler.go:399] "Volume detached for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-trusted-ca-bundle\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.171976 2588 reconciler.go:399] "Volume detached for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-thanos-sidecar-tls\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.171997 2588 reconciler.go:399] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-config-out\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.172013 2588 reconciler.go:399] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-tls-assets\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.172030 2588 reconciler.go:399] "Volume detached for volume \"secret-kube-etcd-client-certs\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-kube-etcd-client-certs\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.172046 2588 reconciler.go:399] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-config\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.172063 2588 reconciler.go:399] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-web-config\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.172082 2588 reconciler.go:399] "Volume detached for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-kubelet-serving-ca-bundle\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.172101 2588 reconciler.go:399] "Volume detached for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-metrics-client-ca\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.172119 2588 reconciler.go:399] "Volume detached for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-metrics-client-certs\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.172157 2588 reconciler.go:399] "Volume detached for volume \"kube-api-access-98gp9\" (UniqueName: \"kubernetes.io/projected/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-kube-api-access-98gp9\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.172175 2588 reconciler.go:399] "Volume detached for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-kube-rbac-proxy\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.172192 2588 reconciler.go:399] "Volume detached for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-grpc-tls\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.172210 2588 reconciler.go:399] "Volume detached for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-configmap-metrics-client-ca\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.172228 2588 reconciler.go:399] "Volume detached for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-secret-prometheus-k8s-tls\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.172245 2588 reconciler.go:399] "Volume detached for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879-prometheus-k8s-rulefiles-0\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.423431 2588 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879/prometheus-proxy/2.log" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.424026 2588 logs.go:323] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879/config-reloader/2.log" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.424667 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event=&{ID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 Type:ContainerDied Data:575fca5742eeb550505988169b65e39889d4e9a4a916a669f68c988db3eb4df1} Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.424705 2588 scope.go:115] "RemoveContainer" containerID="edb583e71afefdc5a9b0a5e0042237321ce66a47c359a57e1eec8611ac5ed37d" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:05.427901747Z" level=info msg="Removing container: edb583e71afefdc5a9b0a5e0042237321ce66a47c359a57e1eec8611ac5ed37d" id=c6de858e-7e9a-484d-95c9-ab397a6c0aa0 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:05.465181568Z" level=info msg="Removed container edb583e71afefdc5a9b0a5e0042237321ce66a47c359a57e1eec8611ac5ed37d: openshift-monitoring/prometheus-k8s-0/kube-rbac-proxy-thanos" id=c6de858e-7e9a-484d-95c9-ab397a6c0aa0 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.465411 2588 scope.go:115] "RemoveContainer" containerID="7de6261b7ebb7d69cfdcbb24569dafb351175e25ae8aecd812d15cab2e644276" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:05.466241908Z" level=info msg="Removing container: 7de6261b7ebb7d69cfdcbb24569dafb351175e25ae8aecd812d15cab2e644276" id=dee80deb-7cce-4136-af99-c75ce3b53c06 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.483198 2588 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-monitoring/prometheus-k8s-0] Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.495102 2588 kubelet.go:2129] "SyncLoop REMOVE" source="api" pods=[openshift-monitoring/prometheus-k8s-0] Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:05.508221953Z" level=info msg="Removed container 7de6261b7ebb7d69cfdcbb24569dafb351175e25ae8aecd812d15cab2e644276: openshift-monitoring/prometheus-k8s-0/kube-rbac-proxy" id=dee80deb-7cce-4136-af99-c75ce3b53c06 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.508859 2588 scope.go:115] "RemoveContainer" containerID="a988dfd069a4d8b4e201e470b1b75848391f749fd271851262309d39a93dc410" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:05.512035342Z" level=info msg="Removing container: a988dfd069a4d8b4e201e470b1b75848391f749fd271851262309d39a93dc410" id=9af49438-590d-4f9a-b9ec-63a5eeb66231 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:05.537529430Z" level=info msg="Removed container a988dfd069a4d8b4e201e470b1b75848391f749fd271851262309d39a93dc410: openshift-monitoring/prometheus-k8s-0/prometheus-proxy" id=9af49438-590d-4f9a-b9ec-63a5eeb66231 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.537873 2588 scope.go:115] "RemoveContainer" containerID="7a2f8f0ad0be821f8959bfd50621160131d618f2b80fdd07c0192537a6788d3a" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:05.538965681Z" level=info msg="Removing container: 7a2f8f0ad0be821f8959bfd50621160131d618f2b80fdd07c0192537a6788d3a" id=3f5b421f-bd2f-4e07-a516-8e0d54d9ada6 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:05.567318945Z" level=info msg="Removed container 7a2f8f0ad0be821f8959bfd50621160131d618f2b80fdd07c0192537a6788d3a: openshift-monitoring/prometheus-k8s-0/thanos-sidecar" id=3f5b421f-bd2f-4e07-a516-8e0d54d9ada6 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.567634 2588 scope.go:115] "RemoveContainer" containerID="bef5095b779c6d727ff2f590bfd7c28dbad80564fc113cb2c36ebaa3ebaeaedf" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:05.571559745Z" level=info msg="Removing container: bef5095b779c6d727ff2f590bfd7c28dbad80564fc113cb2c36ebaa3ebaeaedf" id=af5fd459-fe7d-4505-a3e3-8b2ae89db6b2 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:05.596475724Z" level=info msg="Removed container bef5095b779c6d727ff2f590bfd7c28dbad80564fc113cb2c36ebaa3ebaeaedf: openshift-monitoring/prometheus-k8s-0/config-reloader" id=af5fd459-fe7d-4505-a3e3-8b2ae89db6b2 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.597386 2588 scope.go:115] "RemoveContainer" containerID="4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:05.598807995Z" level=info msg="Removing container: 4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73" id=22108d85-1782-4090-936b-e2672d39ddd5 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:05.621573334Z" level=info msg="Removed container 4a90d2b4d913d98224d46f52ac8b01dee1ae473bbb9689ff7de1041115e24f73: openshift-monitoring/prometheus-k8s-0/prometheus" id=22108d85-1782-4090-936b-e2672d39ddd5 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:05.622594805Z" level=info msg="Removing container: db2619f5536b755153c7510155fc858ed5569d94d503f7db584520ce0158dd84" id=d6f9a60a-4c1d-45a0-8fee-e6d66bff6e17 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:05.621756 2588 scope.go:115] "RemoveContainer" containerID="db2619f5536b755153c7510155fc858ed5569d94d503f7db584520ce0158dd84" Jan 05 09:45:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:05.656723237Z" level=info msg="Removed container db2619f5536b755153c7510155fc858ed5569d94d503f7db584520ce0158dd84: openshift-monitoring/prometheus-k8s-0/init-config-reloader" id=d6f9a60a-4c1d-45a0-8fee-e6d66bff6e17 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:06.215618970Z" level=warning msg="Stopping container 670e670ccf20be64b02c12b1075000c60fafc29d4c8965aee4edaaf717c14ced with stop signal timed out: timeout reached after 30 seconds waiting for container process to exit" id=f56b897a-0771-42f9-9cbb-227a9aee7a07 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:06.382646 2588 cadvisor_stats_provider.go:457] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1166398_2d2c_48bb_8d21_7ca37cefe342.slice/crio-b7acbcee674c440f0bf1b02b910456c58b573f0e935553e9c34a3510945d22ae.scope\": RecentStats: unable to find data in memory cache]" Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:06.397286 2588 cadvisor_stats_provider.go:457] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1166398_2d2c_48bb_8d21_7ca37cefe342.slice/crio-b7acbcee674c440f0bf1b02b910456c58b573f0e935553e9c34a3510945d22ae.scope\": RecentStats: unable to find data in memory cache]" Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:45:06.404297 2588 manager.go:698] Error getting data for container /kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1166398_2d2c_48bb_8d21_7ca37cefe342.slice/crio-b7acbcee674c440f0bf1b02b910456c58b573f0e935553e9c34a3510945d22ae.scope because of race condition Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:06.411786 2588 cadvisor_stats_provider.go:457] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1166398_2d2c_48bb_8d21_7ca37cefe342.slice/crio-b7acbcee674c440f0bf1b02b910456c58b573f0e935553e9c34a3510945d22ae.scope\": RecentStats: unable to find data in memory cache]" Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:45:06.439579 2588 manager.go:698] Error getting data for container /kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05c9a6ea_140a_49a2_bde5_f220b05aa252.slice/crio-conmon-29d242ba69edfb0ca301e890f9169a7355a0889dbda9e344277448360467524b.scope because of race condition Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:45:06.441258 2588 manager.go:698] Error getting data for container /kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9de5de36_9b80_4a0b_a615_f0f9ccda2ff6.slice/crio-conmon-489142621cb0989124915d3c9efa7b65c6ff3af9ed775b03b9f887e6ad7d1d8b.scope because of race condition Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:45:06.441405 2588 manager.go:698] Error getting data for container /kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05c9a6ea_140a_49a2_bde5_f220b05aa252.slice/crio-conmon-4c970eddded2809e2e506cfb0b207a19b375bc49b3a1f9bfd58b202b7c1f6ebd.scope because of race condition Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:45:06.441879 2588 manager.go:698] Error getting data for container /kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c63c9be_d1ad_44b6_8b33_09af3ee314af.slice/crio-578cce0ed1aa4795c3b6d8648ecf4b937233d871e6a027eb233bec20d4333c1f.scope because of race condition Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:45:06.442469 2588 manager.go:698] Error getting data for container /kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c63c9be_d1ad_44b6_8b33_09af3ee314af.slice/crio-conmon-3cbc4c089943552285b7235ee6268116978902d078a993e4f88285e0c16777ce.scope because of race condition Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:45:06.443728 2588 manager.go:698] Error getting data for container /kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ecd2e2c_ab52_42e0_8fde_64e45804cc9e.slice/crio-46e4da5180913795c57fd23e927474161e2562ce8f557a4ef7d187e1da8df07c.scope because of race condition Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:45:06.444432 2588 manager.go:698] Error getting data for container /kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ad4cdcf_3061_4f41_9ab3_5f7f06e2b879.slice/crio-conmon-edb583e71afefdc5a9b0a5e0042237321ce66a47c359a57e1eec8611ac5ed37d.scope because of race condition Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:45:06.444550 2588 manager.go:698] Error getting data for container /kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c63c9be_d1ad_44b6_8b33_09af3ee314af.slice/crio-conmon-c507c8393ab5e7e8bb6dbc97bcabc3f2598889f9b13fbee87614f2fee17b1e9d.scope because of race condition Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:06.452711796Z" level=info msg="Stopped container 670e670ccf20be64b02c12b1075000c60fafc29d4c8965aee4edaaf717c14ced: openshift-multus/cni-sysctl-allowlist-ds-hrwrd/kube-multus-additional-cni-plugins" id=f56b897a-0771-42f9-9cbb-227a9aee7a07 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:06.457579600Z" level=info msg="Stopping pod sandbox: b7acbcee674c440f0bf1b02b910456c58b573f0e935553e9c34a3510945d22ae" id=c7f2d6a0-60b8-48f6-ba0c-e730be1ba90d name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:06.457953481Z" level=info msg="Got pod network &{Name:cni-sysctl-allowlist-ds-hrwrd Namespace:openshift-multus ID:b7acbcee674c440f0bf1b02b910456c58b573f0e935553e9c34a3510945d22ae UID:e1166398-2d2c-48bb-8d21-7ca37cefe342 NetNS:/var/run/netns/6ad09ef5-2f23-47e3-afe2-734fa0458f87 Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:06.458102880Z" level=info msg="Deleting pod openshift-multus_cni-sysctl-allowlist-ds-hrwrd from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:06.461950 2588 generic.go:296] "Generic (PLEG): container finished" podID=e1166398-2d2c-48bb-8d21-7ca37cefe342 containerID="670e670ccf20be64b02c12b1075000c60fafc29d4c8965aee4edaaf717c14ced" exitCode=137 Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:06.461984 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-hrwrd" event=&{ID:e1166398-2d2c-48bb-8d21-7ca37cefe342 Type:ContainerDied Data:670e670ccf20be64b02c12b1075000c60fafc29d4c8965aee4edaaf717c14ced} Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:45:06Z [verbose] Del: openshift-multus:cni-sysctl-allowlist-ds-hrwrd:e1166398-2d2c-48bb-8d21-7ca37cefe342:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:06.709391741Z" level=info msg="Stopped pod sandbox: b7acbcee674c440f0bf1b02b910456c58b573f0e935553e9c34a3510945d22ae" id=c7f2d6a0-60b8-48f6-ba0c-e730be1ba90d name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:06.796061985Z" level=info msg="Stopping pod sandbox: 575fca5742eeb550505988169b65e39889d4e9a4a916a669f68c988db3eb4df1" id=c8e3d24c-83b7-42dc-a851-5e1333d9da21 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:06.796104737Z" level=info msg="Stopped pod sandbox (already stopped): 575fca5742eeb550505988169b65e39889d4e9a4a916a669f68c988db3eb4df1" id=c8e3d24c-83b7-42dc-a851-5e1333d9da21 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:06.798253 2588 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 path="/var/lib/kubelet/pods/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879/volumes" Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:06.830748 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e1166398-2d2c-48bb-8d21-7ca37cefe342-cni-sysctl-allowlist\") pod \"e1166398-2d2c-48bb-8d21-7ca37cefe342\" (UID: \"e1166398-2d2c-48bb-8d21-7ca37cefe342\") " Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:06.830793 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcbvl\" (UniqueName: \"kubernetes.io/projected/e1166398-2d2c-48bb-8d21-7ca37cefe342-kube-api-access-lcbvl\") pod \"e1166398-2d2c-48bb-8d21-7ca37cefe342\" (UID: \"e1166398-2d2c-48bb-8d21-7ca37cefe342\") " Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:06.830830 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/e1166398-2d2c-48bb-8d21-7ca37cefe342-ready\") pod \"e1166398-2d2c-48bb-8d21-7ca37cefe342\" (UID: \"e1166398-2d2c-48bb-8d21-7ca37cefe342\") " Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:06.830863 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e1166398-2d2c-48bb-8d21-7ca37cefe342-tuning-conf-dir\") pod \"e1166398-2d2c-48bb-8d21-7ca37cefe342\" (UID: \"e1166398-2d2c-48bb-8d21-7ca37cefe342\") " Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:06.831051 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1166398-2d2c-48bb-8d21-7ca37cefe342-tuning-conf-dir" (OuterVolumeSpecName: "tuning-conf-dir") pod "e1166398-2d2c-48bb-8d21-7ca37cefe342" (UID: "e1166398-2d2c-48bb-8d21-7ca37cefe342"). InnerVolumeSpecName "tuning-conf-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:06.831175 2588 reconciler.go:399] "Volume detached for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e1166398-2d2c-48bb-8d21-7ca37cefe342-tuning-conf-dir\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:45:06.831518 2588 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/e1166398-2d2c-48bb-8d21-7ca37cefe342/volumes/kubernetes.io~configmap/cni-sysctl-allowlist: clearQuota called, but quotas disabled Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:06.831718 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1166398-2d2c-48bb-8d21-7ca37cefe342-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "e1166398-2d2c-48bb-8d21-7ca37cefe342" (UID: "e1166398-2d2c-48bb-8d21-7ca37cefe342"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:45:06.832558 2588 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/e1166398-2d2c-48bb-8d21-7ca37cefe342/volumes/kubernetes.io~empty-dir/ready: clearQuota called, but quotas disabled Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:06.832648 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1166398-2d2c-48bb-8d21-7ca37cefe342-ready" (OuterVolumeSpecName: "ready") pod "e1166398-2d2c-48bb-8d21-7ca37cefe342" (UID: "e1166398-2d2c-48bb-8d21-7ca37cefe342"). InnerVolumeSpecName "ready". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:06.840609 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1166398-2d2c-48bb-8d21-7ca37cefe342-kube-api-access-lcbvl" (OuterVolumeSpecName: "kube-api-access-lcbvl") pod "e1166398-2d2c-48bb-8d21-7ca37cefe342" (UID: "e1166398-2d2c-48bb-8d21-7ca37cefe342"). InnerVolumeSpecName "kube-api-access-lcbvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:06.932020 2588 reconciler.go:399] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e1166398-2d2c-48bb-8d21-7ca37cefe342-cni-sysctl-allowlist\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:06.932054 2588 reconciler.go:399] "Volume detached for volume \"kube-api-access-lcbvl\" (UniqueName: \"kubernetes.io/projected/e1166398-2d2c-48bb-8d21-7ca37cefe342-kube-api-access-lcbvl\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:06.932069 2588 reconciler.go:399] "Volume detached for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/e1166398-2d2c-48bb-8d21-7ca37cefe342-ready\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:07.465377 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-hrwrd" event=&{ID:e1166398-2d2c-48bb-8d21-7ca37cefe342 Type:ContainerDied Data:b7acbcee674c440f0bf1b02b910456c58b573f0e935553e9c34a3510945d22ae} Jan 05 09:45:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:07.465415 2588 scope.go:115] "RemoveContainer" containerID="670e670ccf20be64b02c12b1075000c60fafc29d4c8965aee4edaaf717c14ced" Jan 05 09:45:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:07.471810832Z" level=info msg="Removing container: 670e670ccf20be64b02c12b1075000c60fafc29d4c8965aee4edaaf717c14ced" id=48a02630-b28d-45a7-91a7-dfd1dbe41196 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:45:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:07.488802 2588 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-multus/cni-sysctl-allowlist-ds-hrwrd] Jan 05 09:45:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:07.493407 2588 kubelet.go:2129] "SyncLoop REMOVE" source="api" pods=[openshift-multus/cni-sysctl-allowlist-ds-hrwrd] Jan 05 09:45:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:07.497284907Z" level=info msg="Removed container 670e670ccf20be64b02c12b1075000c60fafc29d4c8965aee4edaaf717c14ced: openshift-multus/cni-sysctl-allowlist-ds-hrwrd/kube-multus-additional-cni-plugins" id=48a02630-b28d-45a7-91a7-dfd1dbe41196 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:45:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:08.796503 2588 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=e1166398-2d2c-48bb-8d21-7ca37cefe342 path="/var/lib/kubelet/pods/e1166398-2d2c-48bb-8d21-7ca37cefe342/volumes" Jan 05 09:45:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:12.061738 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-image-registry/image-registry-64655f6865-jdhwb" podUID=ba5b2420-eec2-4a15-96cc-fb8dd15e86e8 containerName="registry" containerID="cri-o://85ffd0a49a25ead6a76fd736bb42c420582a8fb24175f37742eccf7af61d0088" gracePeriod=55 Jan 05 09:45:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:12.061995272Z" level=info msg="Stopping container: 85ffd0a49a25ead6a76fd736bb42c420582a8fb24175f37742eccf7af61d0088 (timeout: 55s)" id=bea2c85b-e42a-40b6-9339-d9a2651014dc name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:45:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:12.445448 2588 patch_prober.go:29] interesting pod/router-default-5cf6bbdff9-tnffs container/router namespace/openshift-ingress: Readiness probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]backend-http ok Jan 05 09:45:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: [+]has-synced ok Jan 05 09:45:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: [-]process-running failed: reason withheld Jan 05 09:45:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: healthz check failed Jan 05 09:45:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:12.445505 2588 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" podUID=c3b10ff4-a243-4a4e-b27d-f613e258666d containerName="router" probeResult=failure output="HTTP probe failed with statuscode: 500" Jan 05 09:45:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:12.445575 2588 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" Jan 05 09:45:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:12.625908 2588 patch_prober.go:29] interesting pod/image-registry-64655f6865-jdhwb container/registry namespace/openshift-image-registry: Readiness probe status=failure output="Get \"https://10.131.0.16:5000/healthz\": dial tcp 10.131.0.16:5000: connect: connection refused" start-of-body= Jan 05 09:45:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:12.625966 2588 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-image-registry/image-registry-64655f6865-jdhwb" podUID=ba5b2420-eec2-4a15-96cc-fb8dd15e86e8 containerName="registry" probeResult=failure output="Get \"https://10.131.0.16:5000/healthz\": dial tcp 10.131.0.16:5000: connect: connection refused" Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:13.217442937Z" level=info msg="Stopped container 85ffd0a49a25ead6a76fd736bb42c420582a8fb24175f37742eccf7af61d0088: openshift-image-registry/image-registry-64655f6865-jdhwb/registry" id=bea2c85b-e42a-40b6-9339-d9a2651014dc name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:13.217858766Z" level=info msg="Stopping pod sandbox: 9640eb4f9700129ebcb44b35913e1a9dfbf8801ccc331fb964566aaba63d21d3" id=16e561e6-72d2-47a3-9dd2-24eee5d70b98 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:13.218155613Z" level=info msg="Got pod network &{Name:image-registry-64655f6865-jdhwb Namespace:openshift-image-registry ID:9640eb4f9700129ebcb44b35913e1a9dfbf8801ccc331fb964566aaba63d21d3 UID:ba5b2420-eec2-4a15-96cc-fb8dd15e86e8 NetNS:/var/run/netns/ddd704d5-9de0-48ff-80c5-4a9c024fe20c Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:13.218291189Z" level=info msg="Deleting pod openshift-image-registry_image-registry-64655f6865-jdhwb from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:45:13Z [verbose] Del: openshift-image-registry:image-registry-64655f6865-jdhwb:ba5b2420-eec2-4a15-96cc-fb8dd15e86e8:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:13.360425251Z" level=info msg="Stopped pod sandbox: 9640eb4f9700129ebcb44b35913e1a9dfbf8801ccc331fb964566aaba63d21d3" id=16e561e6-72d2-47a3-9dd2-24eee5d70b98 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.473559 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-registry-tls\") pod \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.473612 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-ca-trust-extracted\") pod \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.473653 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"kube-api-access-484qm\" (UniqueName: \"kubernetes.io/projected/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-kube-api-access-484qm\") pod \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.473688 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-bound-sa-token\") pod \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.473722 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-registry-certificates\") pod \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.473758 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-trusted-ca\") pod \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.473792 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-image-registry-private-configuration\") pod \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.473824 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-installation-pull-secrets\") pod \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\" (UID: \"ba5b2420-eec2-4a15-96cc-fb8dd15e86e8\") " Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:45:13.474726 2588 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8/volumes/kubernetes.io~empty-dir/ca-trust-extracted: clearQuota called, but quotas disabled Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.475370 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "ba5b2420-eec2-4a15-96cc-fb8dd15e86e8" (UID: "ba5b2420-eec2-4a15-96cc-fb8dd15e86e8"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:45:13.475700 2588 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8/volumes/kubernetes.io~configmap/trusted-ca: clearQuota called, but quotas disabled Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.475950 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "ba5b2420-eec2-4a15-96cc-fb8dd15e86e8" (UID: "ba5b2420-eec2-4a15-96cc-fb8dd15e86e8"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:45:13.476093 2588 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8/volumes/kubernetes.io~configmap/registry-certificates: clearQuota called, but quotas disabled Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.476349 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "ba5b2420-eec2-4a15-96cc-fb8dd15e86e8" (UID: "ba5b2420-eec2-4a15-96cc-fb8dd15e86e8"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.483713 2588 generic.go:296] "Generic (PLEG): container finished" podID=ba5b2420-eec2-4a15-96cc-fb8dd15e86e8 containerID="85ffd0a49a25ead6a76fd736bb42c420582a8fb24175f37742eccf7af61d0088" exitCode=0 Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.483751 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-64655f6865-jdhwb" event=&{ID:ba5b2420-eec2-4a15-96cc-fb8dd15e86e8 Type:ContainerDied Data:85ffd0a49a25ead6a76fd736bb42c420582a8fb24175f37742eccf7af61d0088} Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.483778 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-64655f6865-jdhwb" event=&{ID:ba5b2420-eec2-4a15-96cc-fb8dd15e86e8 Type:ContainerDied Data:9640eb4f9700129ebcb44b35913e1a9dfbf8801ccc331fb964566aaba63d21d3} Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.483797 2588 scope.go:115] "RemoveContainer" containerID="85ffd0a49a25ead6a76fd736bb42c420582a8fb24175f37742eccf7af61d0088" Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:13.485102001Z" level=info msg="Removing container: 85ffd0a49a25ead6a76fd736bb42c420582a8fb24175f37742eccf7af61d0088" id=a38572c6-be25-4853-a05a-706d6c7ab163 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.495439 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "ba5b2420-eec2-4a15-96cc-fb8dd15e86e8" (UID: "ba5b2420-eec2-4a15-96cc-fb8dd15e86e8"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.495438 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "ba5b2420-eec2-4a15-96cc-fb8dd15e86e8" (UID: "ba5b2420-eec2-4a15-96cc-fb8dd15e86e8"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.495828 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "ba5b2420-eec2-4a15-96cc-fb8dd15e86e8" (UID: "ba5b2420-eec2-4a15-96cc-fb8dd15e86e8"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.496079 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-kube-api-access-484qm" (OuterVolumeSpecName: "kube-api-access-484qm") pod "ba5b2420-eec2-4a15-96cc-fb8dd15e86e8" (UID: "ba5b2420-eec2-4a15-96cc-fb8dd15e86e8"). InnerVolumeSpecName "kube-api-access-484qm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.497779 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-image-registry-private-configuration" (OuterVolumeSpecName: "image-registry-private-configuration") pod "ba5b2420-eec2-4a15-96cc-fb8dd15e86e8" (UID: "ba5b2420-eec2-4a15-96cc-fb8dd15e86e8"). InnerVolumeSpecName "image-registry-private-configuration". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:13.526687254Z" level=info msg="Removed container 85ffd0a49a25ead6a76fd736bb42c420582a8fb24175f37742eccf7af61d0088: openshift-image-registry/image-registry-64655f6865-jdhwb/registry" id=a38572c6-be25-4853-a05a-706d6c7ab163 name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.528092 2588 scope.go:115] "RemoveContainer" containerID="85ffd0a49a25ead6a76fd736bb42c420582a8fb24175f37742eccf7af61d0088" Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:13.529579 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85ffd0a49a25ead6a76fd736bb42c420582a8fb24175f37742eccf7af61d0088\": container with ID starting with 85ffd0a49a25ead6a76fd736bb42c420582a8fb24175f37742eccf7af61d0088 not found: ID does not exist" containerID="85ffd0a49a25ead6a76fd736bb42c420582a8fb24175f37742eccf7af61d0088" Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.529623 2588 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:85ffd0a49a25ead6a76fd736bb42c420582a8fb24175f37742eccf7af61d0088} err="failed to get container status \"85ffd0a49a25ead6a76fd736bb42c420582a8fb24175f37742eccf7af61d0088\": rpc error: code = NotFound desc = could not find container \"85ffd0a49a25ead6a76fd736bb42c420582a8fb24175f37742eccf7af61d0088\": container with ID starting with 85ffd0a49a25ead6a76fd736bb42c420582a8fb24175f37742eccf7af61d0088 not found: ID does not exist" Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.574970 2588 reconciler.go:399] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-registry-certificates\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.575008 2588 reconciler.go:399] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-bound-sa-token\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.575031 2588 reconciler.go:399] "Volume detached for volume \"image-registry-private-configuration\" (UniqueName: \"kubernetes.io/secret/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-image-registry-private-configuration\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.575049 2588 reconciler.go:399] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-trusted-ca\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.575070 2588 reconciler.go:399] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-installation-pull-secrets\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.575089 2588 reconciler.go:399] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-registry-tls\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.575109 2588 reconciler.go:399] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-ca-trust-extracted\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.575127 2588 reconciler.go:399] "Volume detached for volume \"kube-api-access-484qm\" (UniqueName: \"kubernetes.io/projected/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8-kube-api-access-484qm\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.804616 2588 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-image-registry/image-registry-64655f6865-jdhwb] Jan 05 09:45:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:13.811389 2588 kubelet.go:2129] "SyncLoop REMOVE" source="api" pods=[openshift-image-registry/image-registry-64655f6865-jdhwb] Jan 05 09:45:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:14.794436 2588 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-image-registry/image-registry-64655f6865-jdhwb" podUID=ba5b2420-eec2-4a15-96cc-fb8dd15e86e8 containerName="registry" containerID="cri-o://85ffd0a49a25ead6a76fd736bb42c420582a8fb24175f37742eccf7af61d0088" gracePeriod=1 Jan 05 09:45:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:14.795285810Z" level=info msg="Stopping container: 85ffd0a49a25ead6a76fd736bb42c420582a8fb24175f37742eccf7af61d0088 (timeout: 1s)" id=f6df4980-9e59-401c-9cf7-73792a088311 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:45:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:14.795491 2588 remote_runtime.go:505] "StopContainer from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85ffd0a49a25ead6a76fd736bb42c420582a8fb24175f37742eccf7af61d0088\": container with ID starting with 85ffd0a49a25ead6a76fd736bb42c420582a8fb24175f37742eccf7af61d0088 not found: ID does not exist" containerID="85ffd0a49a25ead6a76fd736bb42c420582a8fb24175f37742eccf7af61d0088" Jan 05 09:45:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:14.795919528Z" level=info msg="Stopping pod sandbox: 9640eb4f9700129ebcb44b35913e1a9dfbf8801ccc331fb964566aaba63d21d3" id=726a601c-f72f-41c9-9811-406d37892bc2 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:14.795964169Z" level=info msg="Stopped pod sandbox (already stopped): 9640eb4f9700129ebcb44b35913e1a9dfbf8801ccc331fb964566aaba63d21d3" id=726a601c-f72f-41c9-9811-406d37892bc2 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:14 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:14.796283 2588 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=ba5b2420-eec2-4a15-96cc-fb8dd15e86e8 path="/var/lib/kubelet/pods/ba5b2420-eec2-4a15-96cc-fb8dd15e86e8/volumes" Jan 05 09:45:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:17.542183897Z" level=info msg="cleanup sandbox network" Jan 05 09:45:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:17.542562179Z" level=info msg="Got pod network &{Name:prometheus-k8s-0 Namespace:openshift-monitoring ID:04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0 UID:4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879 NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:45:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:17.542728610Z" level=info msg="Deleting pod openshift-monitoring_prometheus-k8s-0 from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:45:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:22.444866 2588 patch_prober.go:29] interesting pod/router-default-5cf6bbdff9-tnffs container/router namespace/openshift-ingress: Readiness probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]backend-http ok Jan 05 09:45:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: [+]has-synced ok Jan 05 09:45:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: [-]process-running failed: reason withheld Jan 05 09:45:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: healthz check failed Jan 05 09:45:22 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:22.444931 2588 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" podUID=c3b10ff4-a243-4a4e-b27d-f613e258666d containerName="router" probeResult=failure output="HTTP probe failed with statuscode: 500" Jan 05 09:45:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:32.445111 2588 patch_prober.go:29] interesting pod/router-default-5cf6bbdff9-tnffs container/router namespace/openshift-ingress: Readiness probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 05 09:45:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: [+]has-synced ok Jan 05 09:45:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: [-]process-running failed: reason withheld Jan 05 09:45:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: healthz check failed Jan 05 09:45:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:32.445192 2588 prober.go:114] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" podUID=c3b10ff4-a243-4a4e-b27d-f613e258666d containerName="router" probeResult=failure output="HTTP probe failed with statuscode: 500" Jan 05 09:45:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:32.653612286Z" level=info msg="Stopped container 8eaf1cd2c43d71002a86869a186ab94e90ccc29be62720e01c85c11de8ae9285: openshift-ingress/router-default-5cf6bbdff9-tnffs/router" id=b7fc2463-15a0-46b6-a852-a08443a134ca name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:45:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:32.654195509Z" level=info msg="Stopping pod sandbox: 4814a25d28532df204d5f61d2aaabc14a67b24096511e9cd393f31ea5787ca12" id=94662ba0-cd62-40f6-9834-d8fa7a062800 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:32.654490885Z" level=info msg="Got pod network &{Name:router-default-5cf6bbdff9-tnffs Namespace:openshift-ingress ID:4814a25d28532df204d5f61d2aaabc14a67b24096511e9cd393f31ea5787ca12 UID:c3b10ff4-a243-4a4e-b27d-f613e258666d NetNS:/var/run/netns/d85fb3f0-7148-4415-ac9c-470bac5d14d4 Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:45:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:32.654592573Z" level=info msg="Deleting pod openshift-ingress_router-default-5cf6bbdff9-tnffs from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:45:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: 2023-01-05T09:45:32Z [verbose] Del: openshift-ingress:router-default-5cf6bbdff9-tnffs:c3b10ff4-a243-4a4e-b27d-f613e258666d:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:45:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:32.802430409Z" level=info msg="Stopped pod sandbox: 4814a25d28532df204d5f61d2aaabc14a67b24096511e9cd393f31ea5787ca12" id=94662ba0-cd62-40f6-9834-d8fa7a062800 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:32.991042 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-default-certificate\") pod \"c3b10ff4-a243-4a4e-b27d-f613e258666d\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " Jan 05 09:45:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:32.991096 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3b10ff4-a243-4a4e-b27d-f613e258666d-service-ca-bundle\") pod \"c3b10ff4-a243-4a4e-b27d-f613e258666d\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " Jan 05 09:45:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:32.991125 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ccjjw\" (UniqueName: \"kubernetes.io/projected/c3b10ff4-a243-4a4e-b27d-f613e258666d-kube-api-access-ccjjw\") pod \"c3b10ff4-a243-4a4e-b27d-f613e258666d\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " Jan 05 09:45:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:32.991163 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-metrics-certs\") pod \"c3b10ff4-a243-4a4e-b27d-f613e258666d\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " Jan 05 09:45:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:32.991192 2588 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-stats-auth\") pod \"c3b10ff4-a243-4a4e-b27d-f613e258666d\" (UID: \"c3b10ff4-a243-4a4e-b27d-f613e258666d\") " Jan 05 09:45:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: W0105 09:45:32.991714 2588 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/c3b10ff4-a243-4a4e-b27d-f613e258666d/volumes/kubernetes.io~configmap/service-ca-bundle: clearQuota called, but quotas disabled Jan 05 09:45:32 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:32.991937 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3b10ff4-a243-4a4e-b27d-f613e258666d-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c3b10ff4-a243-4a4e-b27d-f613e258666d" (UID: "c3b10ff4-a243-4a4e-b27d-f613e258666d"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:45:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:33.003435 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c3b10ff4-a243-4a4e-b27d-f613e258666d" (UID: "c3b10ff4-a243-4a4e-b27d-f613e258666d"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:45:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:33.005428 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3b10ff4-a243-4a4e-b27d-f613e258666d-kube-api-access-ccjjw" (OuterVolumeSpecName: "kube-api-access-ccjjw") pod "c3b10ff4-a243-4a4e-b27d-f613e258666d" (UID: "c3b10ff4-a243-4a4e-b27d-f613e258666d"). InnerVolumeSpecName "kube-api-access-ccjjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:45:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:33.007424 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c3b10ff4-a243-4a4e-b27d-f613e258666d" (UID: "c3b10ff4-a243-4a4e-b27d-f613e258666d"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:45:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:33.014434 2588 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c3b10ff4-a243-4a4e-b27d-f613e258666d" (UID: "c3b10ff4-a243-4a4e-b27d-f613e258666d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 05 09:45:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:33.091966 2588 reconciler.go:399] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-default-certificate\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:33.091991 2588 reconciler.go:399] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3b10ff4-a243-4a4e-b27d-f613e258666d-service-ca-bundle\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:33.092009 2588 reconciler.go:399] "Volume detached for volume \"kube-api-access-ccjjw\" (UniqueName: \"kubernetes.io/projected/c3b10ff4-a243-4a4e-b27d-f613e258666d-kube-api-access-ccjjw\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:33.092027 2588 reconciler.go:399] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-metrics-certs\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:33.092043 2588 reconciler.go:399] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c3b10ff4-a243-4a4e-b27d-f613e258666d-stats-auth\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:45:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:33.522064 2588 generic.go:296] "Generic (PLEG): container finished" podID=c3b10ff4-a243-4a4e-b27d-f613e258666d containerID="8eaf1cd2c43d71002a86869a186ab94e90ccc29be62720e01c85c11de8ae9285" exitCode=0 Jan 05 09:45:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:33.522103 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" event=&{ID:c3b10ff4-a243-4a4e-b27d-f613e258666d Type:ContainerDied Data:8eaf1cd2c43d71002a86869a186ab94e90ccc29be62720e01c85c11de8ae9285} Jan 05 09:45:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:33.522128 2588 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5cf6bbdff9-tnffs" event=&{ID:c3b10ff4-a243-4a4e-b27d-f613e258666d Type:ContainerDied Data:4814a25d28532df204d5f61d2aaabc14a67b24096511e9cd393f31ea5787ca12} Jan 05 09:45:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:33.522170 2588 scope.go:115] "RemoveContainer" containerID="8eaf1cd2c43d71002a86869a186ab94e90ccc29be62720e01c85c11de8ae9285" Jan 05 09:45:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:33.525957961Z" level=info msg="Removing container: 8eaf1cd2c43d71002a86869a186ab94e90ccc29be62720e01c85c11de8ae9285" id=e4272ee2-a6f3-48c8-afa3-590f9ec77fbb name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:45:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:33.547428961Z" level=info msg="Removed container 8eaf1cd2c43d71002a86869a186ab94e90ccc29be62720e01c85c11de8ae9285: openshift-ingress/router-default-5cf6bbdff9-tnffs/router" id=e4272ee2-a6f3-48c8-afa3-590f9ec77fbb name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:45:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:33.547631 2588 scope.go:115] "RemoveContainer" containerID="8eaf1cd2c43d71002a86869a186ab94e90ccc29be62720e01c85c11de8ae9285" Jan 05 09:45:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:33.547943 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8eaf1cd2c43d71002a86869a186ab94e90ccc29be62720e01c85c11de8ae9285\": container with ID starting with 8eaf1cd2c43d71002a86869a186ab94e90ccc29be62720e01c85c11de8ae9285 not found: ID does not exist" containerID="8eaf1cd2c43d71002a86869a186ab94e90ccc29be62720e01c85c11de8ae9285" Jan 05 09:45:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:33.547984 2588 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:8eaf1cd2c43d71002a86869a186ab94e90ccc29be62720e01c85c11de8ae9285} err="failed to get container status \"8eaf1cd2c43d71002a86869a186ab94e90ccc29be62720e01c85c11de8ae9285\": rpc error: code = NotFound desc = could not find container \"8eaf1cd2c43d71002a86869a186ab94e90ccc29be62720e01c85c11de8ae9285\": container with ID starting with 8eaf1cd2c43d71002a86869a186ab94e90ccc29be62720e01c85c11de8ae9285 not found: ID does not exist" Jan 05 09:45:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:33.578083 2588 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-ingress/router-default-5cf6bbdff9-tnffs] Jan 05 09:45:33 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:33.578704 2588 kubelet.go:2129] "SyncLoop REMOVE" source="api" pods=[openshift-ingress/router-default-5cf6bbdff9-tnffs] Jan 05 09:45:34 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:34.796695 2588 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=c3b10ff4-a243-4a4e-b27d-f613e258666d path="/var/lib/kubelet/pods/c3b10ff4-a243-4a4e-b27d-f613e258666d/volumes" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.679768552Z" level=info msg="Stopping pod sandbox: fcdbc115e8986f7ba4e1fc9bfa0728fd149a52733e5f0956b91d0b8262b33ca6" id=7ae3cc00-1948-4405-81a3-2fe37d208eca name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.679812129Z" level=info msg="Stopped pod sandbox (already stopped): fcdbc115e8986f7ba4e1fc9bfa0728fd149a52733e5f0956b91d0b8262b33ca6" id=7ae3cc00-1948-4405-81a3-2fe37d208eca name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.680437070Z" level=info msg="Removing pod sandbox: fcdbc115e8986f7ba4e1fc9bfa0728fd149a52733e5f0956b91d0b8262b33ca6" id=744374a8-7afe-4fe3-96d3-157a79851f14 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.687199363Z" level=info msg="Removed pod sandbox: fcdbc115e8986f7ba4e1fc9bfa0728fd149a52733e5f0956b91d0b8262b33ca6" id=744374a8-7afe-4fe3-96d3-157a79851f14 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.687569313Z" level=info msg="Stopping pod sandbox: d6d0c82bc9774e51559f70121d26a57c7c38cce6fb83f23b63a8bb671c408015" id=cce283c4-f8cd-483b-b989-86ba6ec97747 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.687597479Z" level=info msg="Stopped pod sandbox (already stopped): d6d0c82bc9774e51559f70121d26a57c7c38cce6fb83f23b63a8bb671c408015" id=cce283c4-f8cd-483b-b989-86ba6ec97747 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.687803001Z" level=info msg="Removing pod sandbox: d6d0c82bc9774e51559f70121d26a57c7c38cce6fb83f23b63a8bb671c408015" id=0004285f-b2ef-4d93-89ae-e6c0e92cf89a name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.694687477Z" level=info msg="Removed pod sandbox: d6d0c82bc9774e51559f70121d26a57c7c38cce6fb83f23b63a8bb671c408015" id=0004285f-b2ef-4d93-89ae-e6c0e92cf89a name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.694974890Z" level=info msg="Stopping pod sandbox: b7acbcee674c440f0bf1b02b910456c58b573f0e935553e9c34a3510945d22ae" id=db9677f8-0811-4033-a334-55ba698f0721 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.695002767Z" level=info msg="Stopped pod sandbox (already stopped): b7acbcee674c440f0bf1b02b910456c58b573f0e935553e9c34a3510945d22ae" id=db9677f8-0811-4033-a334-55ba698f0721 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.695236193Z" level=info msg="Removing pod sandbox: b7acbcee674c440f0bf1b02b910456c58b573f0e935553e9c34a3510945d22ae" id=381bee50-2ddf-4228-9159-89ab6d2fe17a name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.701349634Z" level=info msg="Removed pod sandbox: b7acbcee674c440f0bf1b02b910456c58b573f0e935553e9c34a3510945d22ae" id=381bee50-2ddf-4228-9159-89ab6d2fe17a name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.701633380Z" level=info msg="Stopping pod sandbox: 9640eb4f9700129ebcb44b35913e1a9dfbf8801ccc331fb964566aaba63d21d3" id=697a7c2e-1d72-490c-ace2-2ccd72271f89 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.701666156Z" level=info msg="Stopped pod sandbox (already stopped): 9640eb4f9700129ebcb44b35913e1a9dfbf8801ccc331fb964566aaba63d21d3" id=697a7c2e-1d72-490c-ace2-2ccd72271f89 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.701861396Z" level=info msg="Removing pod sandbox: 9640eb4f9700129ebcb44b35913e1a9dfbf8801ccc331fb964566aaba63d21d3" id=69087c7d-8c01-48f2-82ac-1dc65345a2b6 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.708629043Z" level=info msg="Removed pod sandbox: 9640eb4f9700129ebcb44b35913e1a9dfbf8801ccc331fb964566aaba63d21d3" id=69087c7d-8c01-48f2-82ac-1dc65345a2b6 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.708979975Z" level=info msg="Stopping pod sandbox: 1e97b2e64ee6d4f76d829e9619da041f24e571e3de5454c48314d411b8770c02" id=3f6316a6-8c17-4c69-acc5-194763d86141 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.709014999Z" level=info msg="Stopped pod sandbox (already stopped): 1e97b2e64ee6d4f76d829e9619da041f24e571e3de5454c48314d411b8770c02" id=3f6316a6-8c17-4c69-acc5-194763d86141 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.709288453Z" level=info msg="Removing pod sandbox: 1e97b2e64ee6d4f76d829e9619da041f24e571e3de5454c48314d411b8770c02" id=0493f9d4-aabe-420a-8dc1-7a08b3c38310 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.754218172Z" level=info msg="Removed pod sandbox: 1e97b2e64ee6d4f76d829e9619da041f24e571e3de5454c48314d411b8770c02" id=0493f9d4-aabe-420a-8dc1-7a08b3c38310 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.755123233Z" level=info msg="Stopping pod sandbox: e10e02d14faf7624d78743e0d1899ddf64963db130ecf653b0849bcde995a72c" id=b4fe43cb-bc69-4b5b-b73c-bb4d019eadfc name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.755176659Z" level=info msg="Stopped pod sandbox (already stopped): e10e02d14faf7624d78743e0d1899ddf64963db130ecf653b0849bcde995a72c" id=b4fe43cb-bc69-4b5b-b73c-bb4d019eadfc name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.755531530Z" level=info msg="Removing pod sandbox: e10e02d14faf7624d78743e0d1899ddf64963db130ecf653b0849bcde995a72c" id=632e2f6d-2800-4555-9ab4-665b538c35d6 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.761658169Z" level=info msg="Removed pod sandbox: e10e02d14faf7624d78743e0d1899ddf64963db130ecf653b0849bcde995a72c" id=632e2f6d-2800-4555-9ab4-665b538c35d6 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.761901539Z" level=info msg="Stopping pod sandbox: 575fca5742eeb550505988169b65e39889d4e9a4a916a669f68c988db3eb4df1" id=84464848-a603-47bc-8510-c98e98eff846 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.761926264Z" level=info msg="Stopped pod sandbox (already stopped): 575fca5742eeb550505988169b65e39889d4e9a4a916a669f68c988db3eb4df1" id=84464848-a603-47bc-8510-c98e98eff846 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.762204609Z" level=info msg="Removing pod sandbox: 575fca5742eeb550505988169b65e39889d4e9a4a916a669f68c988db3eb4df1" id=f2f94849-4453-4cbb-ba91-086fd8eb0112 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.774710227Z" level=info msg="Removed pod sandbox: 575fca5742eeb550505988169b65e39889d4e9a4a916a669f68c988db3eb4df1" id=f2f94849-4453-4cbb-ba91-086fd8eb0112 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.774983484Z" level=info msg="Stopping pod sandbox: 9d9807a6998df2507a4aa5b9bad79e347bc60d62b254989738c927929d827e6e" id=17017c29-c481-4b5e-aa5e-e808eee173f6 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.775007084Z" level=info msg="Stopped pod sandbox (already stopped): 9d9807a6998df2507a4aa5b9bad79e347bc60d62b254989738c927929d827e6e" id=17017c29-c481-4b5e-aa5e-e808eee173f6 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.775242322Z" level=info msg="Removing pod sandbox: 9d9807a6998df2507a4aa5b9bad79e347bc60d62b254989738c927929d827e6e" id=8dd087e5-9706-48c5-89f4-17f1329212b3 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.781704370Z" level=info msg="Removed pod sandbox: 9d9807a6998df2507a4aa5b9bad79e347bc60d62b254989738c927929d827e6e" id=8dd087e5-9706-48c5-89f4-17f1329212b3 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.781929449Z" level=info msg="Stopping pod sandbox: f9cc29d007bd3c948c46f781abd08afdb46359ec50fc31f8c503fcdb35b7aa22" id=bec1a812-ceb0-433f-96fa-26b960f9ebb1 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.781955687Z" level=info msg="Stopped pod sandbox (already stopped): f9cc29d007bd3c948c46f781abd08afdb46359ec50fc31f8c503fcdb35b7aa22" id=bec1a812-ceb0-433f-96fa-26b960f9ebb1 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.782190976Z" level=info msg="Removing pod sandbox: f9cc29d007bd3c948c46f781abd08afdb46359ec50fc31f8c503fcdb35b7aa22" id=6c5ba0cb-a839-45f2-9a95-bc99974cebcc name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.788184740Z" level=info msg="Removed pod sandbox: f9cc29d007bd3c948c46f781abd08afdb46359ec50fc31f8c503fcdb35b7aa22" id=6c5ba0cb-a839-45f2-9a95-bc99974cebcc name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.788418351Z" level=info msg="Stopping pod sandbox: 943df75e8c120cec9d37e4a8dd05aed3e8f310e67c6a2c33da550e6cdbd653f4" id=b2dd4863-4d53-4869-8f33-19c2df6cc928 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.788441044Z" level=info msg="Stopped pod sandbox (already stopped): 943df75e8c120cec9d37e4a8dd05aed3e8f310e67c6a2c33da550e6cdbd653f4" id=b2dd4863-4d53-4869-8f33-19c2df6cc928 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.788643782Z" level=info msg="Removing pod sandbox: 943df75e8c120cec9d37e4a8dd05aed3e8f310e67c6a2c33da550e6cdbd653f4" id=4669bea2-f91d-4ac1-a773-8baf0be9d559 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.796540507Z" level=info msg="Removed pod sandbox: 943df75e8c120cec9d37e4a8dd05aed3e8f310e67c6a2c33da550e6cdbd653f4" id=4669bea2-f91d-4ac1-a773-8baf0be9d559 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.796751467Z" level=info msg="Stopping pod sandbox: 7a37710cfebf59892a4d6b3c2e12765b60414f6a7cfb8933a050df609c94e690" id=7168c94c-9253-487a-8bf2-903edfe910b7 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.796779509Z" level=info msg="Stopped pod sandbox (already stopped): 7a37710cfebf59892a4d6b3c2e12765b60414f6a7cfb8933a050df609c94e690" id=7168c94c-9253-487a-8bf2-903edfe910b7 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.797016576Z" level=info msg="Removing pod sandbox: 7a37710cfebf59892a4d6b3c2e12765b60414f6a7cfb8933a050df609c94e690" id=7aac8946-4ddb-4b63-99b7-09b614fb5222 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.803487702Z" level=info msg="Removed pod sandbox: 7a37710cfebf59892a4d6b3c2e12765b60414f6a7cfb8933a050df609c94e690" id=7aac8946-4ddb-4b63-99b7-09b614fb5222 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.803712946Z" level=info msg="Stopping pod sandbox: 4814a25d28532df204d5f61d2aaabc14a67b24096511e9cd393f31ea5787ca12" id=958f3759-a6f3-43f9-b835-28e45c17ba1d name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.803736212Z" level=info msg="Stopped pod sandbox (already stopped): 4814a25d28532df204d5f61d2aaabc14a67b24096511e9cd393f31ea5787ca12" id=958f3759-a6f3-43f9-b835-28e45c17ba1d name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.803921454Z" level=info msg="Removing pod sandbox: 4814a25d28532df204d5f61d2aaabc14a67b24096511e9cd393f31ea5787ca12" id=3fa9f0bc-6858-4071-8b44-74e225b6d3a3 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:45:40.809657768Z" level=info msg="Removed pod sandbox: 4814a25d28532df204d5f61d2aaabc14a67b24096511e9cd393f31ea5787ca12" id=3fa9f0bc-6858-4071-8b44-74e225b6d3a3 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.813453 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"006114a568589213245ba8263ebfa77e665fa4cac909846666c0dede6b7512f9\": container with ID starting with 006114a568589213245ba8263ebfa77e665fa4cac909846666c0dede6b7512f9 not found: ID does not exist" containerID="006114a568589213245ba8263ebfa77e665fa4cac909846666c0dede6b7512f9" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.813494 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="006114a568589213245ba8263ebfa77e665fa4cac909846666c0dede6b7512f9" err="rpc error: code = NotFound desc = could not find container \"006114a568589213245ba8263ebfa77e665fa4cac909846666c0dede6b7512f9\": container with ID starting with 006114a568589213245ba8263ebfa77e665fa4cac909846666c0dede6b7512f9 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.813790 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d7374a8e3df4cc930990f545fee87038e1b2ad2278644240152875a9273e862\": container with ID starting with 3d7374a8e3df4cc930990f545fee87038e1b2ad2278644240152875a9273e862 not found: ID does not exist" containerID="3d7374a8e3df4cc930990f545fee87038e1b2ad2278644240152875a9273e862" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.813823 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="3d7374a8e3df4cc930990f545fee87038e1b2ad2278644240152875a9273e862" err="rpc error: code = NotFound desc = could not find container \"3d7374a8e3df4cc930990f545fee87038e1b2ad2278644240152875a9273e862\": container with ID starting with 3d7374a8e3df4cc930990f545fee87038e1b2ad2278644240152875a9273e862 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.814071 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c1ff2a5effcdc52ca7d965b85b5b36131e77e34ca25d1e8807eedb4eb9c99d0\": container with ID starting with 1c1ff2a5effcdc52ca7d965b85b5b36131e77e34ca25d1e8807eedb4eb9c99d0 not found: ID does not exist" containerID="1c1ff2a5effcdc52ca7d965b85b5b36131e77e34ca25d1e8807eedb4eb9c99d0" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.814098 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="1c1ff2a5effcdc52ca7d965b85b5b36131e77e34ca25d1e8807eedb4eb9c99d0" err="rpc error: code = NotFound desc = could not find container \"1c1ff2a5effcdc52ca7d965b85b5b36131e77e34ca25d1e8807eedb4eb9c99d0\": container with ID starting with 1c1ff2a5effcdc52ca7d965b85b5b36131e77e34ca25d1e8807eedb4eb9c99d0 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.814445 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43228eea868ea1a6b603b81a7625cda951823b52f39f51806c1ba0ccdccbda56\": container with ID starting with 43228eea868ea1a6b603b81a7625cda951823b52f39f51806c1ba0ccdccbda56 not found: ID does not exist" containerID="43228eea868ea1a6b603b81a7625cda951823b52f39f51806c1ba0ccdccbda56" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.814474 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="43228eea868ea1a6b603b81a7625cda951823b52f39f51806c1ba0ccdccbda56" err="rpc error: code = NotFound desc = could not find container \"43228eea868ea1a6b603b81a7625cda951823b52f39f51806c1ba0ccdccbda56\": container with ID starting with 43228eea868ea1a6b603b81a7625cda951823b52f39f51806c1ba0ccdccbda56 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.814717 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65d2017d651f03a83c1df13113775d03f51374c0b6714011dd6c4783e504bbc7\": container with ID starting with 65d2017d651f03a83c1df13113775d03f51374c0b6714011dd6c4783e504bbc7 not found: ID does not exist" containerID="65d2017d651f03a83c1df13113775d03f51374c0b6714011dd6c4783e504bbc7" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.814744 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="65d2017d651f03a83c1df13113775d03f51374c0b6714011dd6c4783e504bbc7" err="rpc error: code = NotFound desc = could not find container \"65d2017d651f03a83c1df13113775d03f51374c0b6714011dd6c4783e504bbc7\": container with ID starting with 65d2017d651f03a83c1df13113775d03f51374c0b6714011dd6c4783e504bbc7 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.814995 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"744edbc686210b5388467d89b22a8f7cad5ce589c1225d191c8e1a4a5f5b4413\": container with ID starting with 744edbc686210b5388467d89b22a8f7cad5ce589c1225d191c8e1a4a5f5b4413 not found: ID does not exist" containerID="744edbc686210b5388467d89b22a8f7cad5ce589c1225d191c8e1a4a5f5b4413" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.815021 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="744edbc686210b5388467d89b22a8f7cad5ce589c1225d191c8e1a4a5f5b4413" err="rpc error: code = NotFound desc = could not find container \"744edbc686210b5388467d89b22a8f7cad5ce589c1225d191c8e1a4a5f5b4413\": container with ID starting with 744edbc686210b5388467d89b22a8f7cad5ce589c1225d191c8e1a4a5f5b4413 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.815282 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7750d5fb0eed5aadcfb300d954cd8a06c38d92001a4babb24681e14257ee1005\": container with ID starting with 7750d5fb0eed5aadcfb300d954cd8a06c38d92001a4babb24681e14257ee1005 not found: ID does not exist" containerID="7750d5fb0eed5aadcfb300d954cd8a06c38d92001a4babb24681e14257ee1005" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.815310 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="7750d5fb0eed5aadcfb300d954cd8a06c38d92001a4babb24681e14257ee1005" err="rpc error: code = NotFound desc = could not find container \"7750d5fb0eed5aadcfb300d954cd8a06c38d92001a4babb24681e14257ee1005\": container with ID starting with 7750d5fb0eed5aadcfb300d954cd8a06c38d92001a4babb24681e14257ee1005 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.815566 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbba8d600bd026ca0a23b1c9724088c18597af8b742d44a66208c8f6b0dbe0ad\": container with ID starting with fbba8d600bd026ca0a23b1c9724088c18597af8b742d44a66208c8f6b0dbe0ad not found: ID does not exist" containerID="fbba8d600bd026ca0a23b1c9724088c18597af8b742d44a66208c8f6b0dbe0ad" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.815593 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="fbba8d600bd026ca0a23b1c9724088c18597af8b742d44a66208c8f6b0dbe0ad" err="rpc error: code = NotFound desc = could not find container \"fbba8d600bd026ca0a23b1c9724088c18597af8b742d44a66208c8f6b0dbe0ad\": container with ID starting with fbba8d600bd026ca0a23b1c9724088c18597af8b742d44a66208c8f6b0dbe0ad not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.815841 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9fdb00607de8750f6e02b9aad69c1eb2f10b013ec69caa09bea48d801e3775d\": container with ID starting with a9fdb00607de8750f6e02b9aad69c1eb2f10b013ec69caa09bea48d801e3775d not found: ID does not exist" containerID="a9fdb00607de8750f6e02b9aad69c1eb2f10b013ec69caa09bea48d801e3775d" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.815868 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="a9fdb00607de8750f6e02b9aad69c1eb2f10b013ec69caa09bea48d801e3775d" err="rpc error: code = NotFound desc = could not find container \"a9fdb00607de8750f6e02b9aad69c1eb2f10b013ec69caa09bea48d801e3775d\": container with ID starting with a9fdb00607de8750f6e02b9aad69c1eb2f10b013ec69caa09bea48d801e3775d not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.816115 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb7fc5ba8ff324ca58d6a410bedf58010254481342f59413cbb3fa0f445a6348\": container with ID starting with eb7fc5ba8ff324ca58d6a410bedf58010254481342f59413cbb3fa0f445a6348 not found: ID does not exist" containerID="eb7fc5ba8ff324ca58d6a410bedf58010254481342f59413cbb3fa0f445a6348" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.816161 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="eb7fc5ba8ff324ca58d6a410bedf58010254481342f59413cbb3fa0f445a6348" err="rpc error: code = NotFound desc = could not find container \"eb7fc5ba8ff324ca58d6a410bedf58010254481342f59413cbb3fa0f445a6348\": container with ID starting with eb7fc5ba8ff324ca58d6a410bedf58010254481342f59413cbb3fa0f445a6348 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.816429 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"136db9dd8f79fe31a8bcb24721d70a69673af9677342ce64467024c361304f6e\": container with ID starting with 136db9dd8f79fe31a8bcb24721d70a69673af9677342ce64467024c361304f6e not found: ID does not exist" containerID="136db9dd8f79fe31a8bcb24721d70a69673af9677342ce64467024c361304f6e" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.816458 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="136db9dd8f79fe31a8bcb24721d70a69673af9677342ce64467024c361304f6e" err="rpc error: code = NotFound desc = could not find container \"136db9dd8f79fe31a8bcb24721d70a69673af9677342ce64467024c361304f6e\": container with ID starting with 136db9dd8f79fe31a8bcb24721d70a69673af9677342ce64467024c361304f6e not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.816763 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4fe2d6c3cc9d0c0aa0e03b9983f735dd906f6abbf9c60a8e821236675195f71\": container with ID starting with f4fe2d6c3cc9d0c0aa0e03b9983f735dd906f6abbf9c60a8e821236675195f71 not found: ID does not exist" containerID="f4fe2d6c3cc9d0c0aa0e03b9983f735dd906f6abbf9c60a8e821236675195f71" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.816791 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="f4fe2d6c3cc9d0c0aa0e03b9983f735dd906f6abbf9c60a8e821236675195f71" err="rpc error: code = NotFound desc = could not find container \"f4fe2d6c3cc9d0c0aa0e03b9983f735dd906f6abbf9c60a8e821236675195f71\": container with ID starting with f4fe2d6c3cc9d0c0aa0e03b9983f735dd906f6abbf9c60a8e821236675195f71 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.817186 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"509279176d1ba7a99db50371d2e6fbd2aa072937e3210855121bfc88ffaf8250\": container with ID starting with 509279176d1ba7a99db50371d2e6fbd2aa072937e3210855121bfc88ffaf8250 not found: ID does not exist" containerID="509279176d1ba7a99db50371d2e6fbd2aa072937e3210855121bfc88ffaf8250" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.817216 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="509279176d1ba7a99db50371d2e6fbd2aa072937e3210855121bfc88ffaf8250" err="rpc error: code = NotFound desc = could not find container \"509279176d1ba7a99db50371d2e6fbd2aa072937e3210855121bfc88ffaf8250\": container with ID starting with 509279176d1ba7a99db50371d2e6fbd2aa072937e3210855121bfc88ffaf8250 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.817470 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af38b04d912e1eb5b634a533910f05eb846127a534a3ac450ce9b629df1b1f1f\": container with ID starting with af38b04d912e1eb5b634a533910f05eb846127a534a3ac450ce9b629df1b1f1f not found: ID does not exist" containerID="af38b04d912e1eb5b634a533910f05eb846127a534a3ac450ce9b629df1b1f1f" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.817497 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="af38b04d912e1eb5b634a533910f05eb846127a534a3ac450ce9b629df1b1f1f" err="rpc error: code = NotFound desc = could not find container \"af38b04d912e1eb5b634a533910f05eb846127a534a3ac450ce9b629df1b1f1f\": container with ID starting with af38b04d912e1eb5b634a533910f05eb846127a534a3ac450ce9b629df1b1f1f not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.817729 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"047cddfaab1d98f2f4be03b086b5acc263c7285299eed9b70a12b9c6485a6435\": container with ID starting with 047cddfaab1d98f2f4be03b086b5acc263c7285299eed9b70a12b9c6485a6435 not found: ID does not exist" containerID="047cddfaab1d98f2f4be03b086b5acc263c7285299eed9b70a12b9c6485a6435" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.817756 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="047cddfaab1d98f2f4be03b086b5acc263c7285299eed9b70a12b9c6485a6435" err="rpc error: code = NotFound desc = could not find container \"047cddfaab1d98f2f4be03b086b5acc263c7285299eed9b70a12b9c6485a6435\": container with ID starting with 047cddfaab1d98f2f4be03b086b5acc263c7285299eed9b70a12b9c6485a6435 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.817998 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2dca518d6d28aeafb7a13c3ccc01e3f5af14216c4354b60ce83f720c1064b4b7\": container with ID starting with 2dca518d6d28aeafb7a13c3ccc01e3f5af14216c4354b60ce83f720c1064b4b7 not found: ID does not exist" containerID="2dca518d6d28aeafb7a13c3ccc01e3f5af14216c4354b60ce83f720c1064b4b7" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.818023 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="2dca518d6d28aeafb7a13c3ccc01e3f5af14216c4354b60ce83f720c1064b4b7" err="rpc error: code = NotFound desc = could not find container \"2dca518d6d28aeafb7a13c3ccc01e3f5af14216c4354b60ce83f720c1064b4b7\": container with ID starting with 2dca518d6d28aeafb7a13c3ccc01e3f5af14216c4354b60ce83f720c1064b4b7 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.818309 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c70e69ff06eb3aacb6dd1829475cbd0126780b7fa69a099ad7ca98bac92ddef\": container with ID starting with 9c70e69ff06eb3aacb6dd1829475cbd0126780b7fa69a099ad7ca98bac92ddef not found: ID does not exist" containerID="9c70e69ff06eb3aacb6dd1829475cbd0126780b7fa69a099ad7ca98bac92ddef" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.818337 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="9c70e69ff06eb3aacb6dd1829475cbd0126780b7fa69a099ad7ca98bac92ddef" err="rpc error: code = NotFound desc = could not find container \"9c70e69ff06eb3aacb6dd1829475cbd0126780b7fa69a099ad7ca98bac92ddef\": container with ID starting with 9c70e69ff06eb3aacb6dd1829475cbd0126780b7fa69a099ad7ca98bac92ddef not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.818571 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a64a8eec7c21e1dc010077a81f71bfd1ef40287e87f8d9147af65ba664e54322\": container with ID starting with a64a8eec7c21e1dc010077a81f71bfd1ef40287e87f8d9147af65ba664e54322 not found: ID does not exist" containerID="a64a8eec7c21e1dc010077a81f71bfd1ef40287e87f8d9147af65ba664e54322" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.818597 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="a64a8eec7c21e1dc010077a81f71bfd1ef40287e87f8d9147af65ba664e54322" err="rpc error: code = NotFound desc = could not find container \"a64a8eec7c21e1dc010077a81f71bfd1ef40287e87f8d9147af65ba664e54322\": container with ID starting with a64a8eec7c21e1dc010077a81f71bfd1ef40287e87f8d9147af65ba664e54322 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.819189 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"011acc34c572433812691a69859a1ec4d3706557a1b172add408479127957e6d\": container with ID starting with 011acc34c572433812691a69859a1ec4d3706557a1b172add408479127957e6d not found: ID does not exist" containerID="011acc34c572433812691a69859a1ec4d3706557a1b172add408479127957e6d" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.819219 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="011acc34c572433812691a69859a1ec4d3706557a1b172add408479127957e6d" err="rpc error: code = NotFound desc = could not find container \"011acc34c572433812691a69859a1ec4d3706557a1b172add408479127957e6d\": container with ID starting with 011acc34c572433812691a69859a1ec4d3706557a1b172add408479127957e6d not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.819477 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a92439f77001fbb754b6894a21e8f5cc53e532d705ed818cd99da13a6465a6ef\": container with ID starting with a92439f77001fbb754b6894a21e8f5cc53e532d705ed818cd99da13a6465a6ef not found: ID does not exist" containerID="a92439f77001fbb754b6894a21e8f5cc53e532d705ed818cd99da13a6465a6ef" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.819507 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="a92439f77001fbb754b6894a21e8f5cc53e532d705ed818cd99da13a6465a6ef" err="rpc error: code = NotFound desc = could not find container \"a92439f77001fbb754b6894a21e8f5cc53e532d705ed818cd99da13a6465a6ef\": container with ID starting with a92439f77001fbb754b6894a21e8f5cc53e532d705ed818cd99da13a6465a6ef not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.819746 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d6d515b36d97644368b6561572ffcff9d150ef935eef5e962c5fabb98e5369a\": container with ID starting with 1d6d515b36d97644368b6561572ffcff9d150ef935eef5e962c5fabb98e5369a not found: ID does not exist" containerID="1d6d515b36d97644368b6561572ffcff9d150ef935eef5e962c5fabb98e5369a" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.819772 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="1d6d515b36d97644368b6561572ffcff9d150ef935eef5e962c5fabb98e5369a" err="rpc error: code = NotFound desc = could not find container \"1d6d515b36d97644368b6561572ffcff9d150ef935eef5e962c5fabb98e5369a\": container with ID starting with 1d6d515b36d97644368b6561572ffcff9d150ef935eef5e962c5fabb98e5369a not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.820010 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"972cda72e60cf7a211036a270be6909a83c8720dc668efa2e2dd303e21789191\": container with ID starting with 972cda72e60cf7a211036a270be6909a83c8720dc668efa2e2dd303e21789191 not found: ID does not exist" containerID="972cda72e60cf7a211036a270be6909a83c8720dc668efa2e2dd303e21789191" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.820039 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="972cda72e60cf7a211036a270be6909a83c8720dc668efa2e2dd303e21789191" err="rpc error: code = NotFound desc = could not find container \"972cda72e60cf7a211036a270be6909a83c8720dc668efa2e2dd303e21789191\": container with ID starting with 972cda72e60cf7a211036a270be6909a83c8720dc668efa2e2dd303e21789191 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.820314 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30b2d68c19c82a725bb94bba811fe395846878cd0d20034d15397fe68fbe2865\": container with ID starting with 30b2d68c19c82a725bb94bba811fe395846878cd0d20034d15397fe68fbe2865 not found: ID does not exist" containerID="30b2d68c19c82a725bb94bba811fe395846878cd0d20034d15397fe68fbe2865" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.820342 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="30b2d68c19c82a725bb94bba811fe395846878cd0d20034d15397fe68fbe2865" err="rpc error: code = NotFound desc = could not find container \"30b2d68c19c82a725bb94bba811fe395846878cd0d20034d15397fe68fbe2865\": container with ID starting with 30b2d68c19c82a725bb94bba811fe395846878cd0d20034d15397fe68fbe2865 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.820576 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1800c468fe04ef1ee6fc5ecf08f9b47946f276d072facfdc0f9944871c84001\": container with ID starting with b1800c468fe04ef1ee6fc5ecf08f9b47946f276d072facfdc0f9944871c84001 not found: ID does not exist" containerID="b1800c468fe04ef1ee6fc5ecf08f9b47946f276d072facfdc0f9944871c84001" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.820604 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="b1800c468fe04ef1ee6fc5ecf08f9b47946f276d072facfdc0f9944871c84001" err="rpc error: code = NotFound desc = could not find container \"b1800c468fe04ef1ee6fc5ecf08f9b47946f276d072facfdc0f9944871c84001\": container with ID starting with b1800c468fe04ef1ee6fc5ecf08f9b47946f276d072facfdc0f9944871c84001 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.820837 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75e3e62829960934d9fc0efc4aa383fbc6135b8a25d451f9e95cc3ccb510acad\": container with ID starting with 75e3e62829960934d9fc0efc4aa383fbc6135b8a25d451f9e95cc3ccb510acad not found: ID does not exist" containerID="75e3e62829960934d9fc0efc4aa383fbc6135b8a25d451f9e95cc3ccb510acad" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.820862 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="75e3e62829960934d9fc0efc4aa383fbc6135b8a25d451f9e95cc3ccb510acad" err="rpc error: code = NotFound desc = could not find container \"75e3e62829960934d9fc0efc4aa383fbc6135b8a25d451f9e95cc3ccb510acad\": container with ID starting with 75e3e62829960934d9fc0efc4aa383fbc6135b8a25d451f9e95cc3ccb510acad not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.821099 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff8e0555d55e91975ea01b26878c5b45f1de8d8d28aa191358f2ab55fccbc19b\": container with ID starting with ff8e0555d55e91975ea01b26878c5b45f1de8d8d28aa191358f2ab55fccbc19b not found: ID does not exist" containerID="ff8e0555d55e91975ea01b26878c5b45f1de8d8d28aa191358f2ab55fccbc19b" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.821125 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="ff8e0555d55e91975ea01b26878c5b45f1de8d8d28aa191358f2ab55fccbc19b" err="rpc error: code = NotFound desc = could not find container \"ff8e0555d55e91975ea01b26878c5b45f1de8d8d28aa191358f2ab55fccbc19b\": container with ID starting with ff8e0555d55e91975ea01b26878c5b45f1de8d8d28aa191358f2ab55fccbc19b not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.821399 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3676f1603de80edd23f34248eef289ffee599487f3f97ddd83f3362b1b7dc2f0\": container with ID starting with 3676f1603de80edd23f34248eef289ffee599487f3f97ddd83f3362b1b7dc2f0 not found: ID does not exist" containerID="3676f1603de80edd23f34248eef289ffee599487f3f97ddd83f3362b1b7dc2f0" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.821426 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="3676f1603de80edd23f34248eef289ffee599487f3f97ddd83f3362b1b7dc2f0" err="rpc error: code = NotFound desc = could not find container \"3676f1603de80edd23f34248eef289ffee599487f3f97ddd83f3362b1b7dc2f0\": container with ID starting with 3676f1603de80edd23f34248eef289ffee599487f3f97ddd83f3362b1b7dc2f0 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.821661 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5456d8a84c429fbed32ddf7f23ca6a6e2863d0fc67333c5d2c2fac4f1a3329e0\": container with ID starting with 5456d8a84c429fbed32ddf7f23ca6a6e2863d0fc67333c5d2c2fac4f1a3329e0 not found: ID does not exist" containerID="5456d8a84c429fbed32ddf7f23ca6a6e2863d0fc67333c5d2c2fac4f1a3329e0" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.821688 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="5456d8a84c429fbed32ddf7f23ca6a6e2863d0fc67333c5d2c2fac4f1a3329e0" err="rpc error: code = NotFound desc = could not find container \"5456d8a84c429fbed32ddf7f23ca6a6e2863d0fc67333c5d2c2fac4f1a3329e0\": container with ID starting with 5456d8a84c429fbed32ddf7f23ca6a6e2863d0fc67333c5d2c2fac4f1a3329e0 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.821923 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87856f0f10585dd547492ed5711f26f3409ec28f07c808cd58bb845c8f6d7522\": container with ID starting with 87856f0f10585dd547492ed5711f26f3409ec28f07c808cd58bb845c8f6d7522 not found: ID does not exist" containerID="87856f0f10585dd547492ed5711f26f3409ec28f07c808cd58bb845c8f6d7522" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.821950 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="87856f0f10585dd547492ed5711f26f3409ec28f07c808cd58bb845c8f6d7522" err="rpc error: code = NotFound desc = could not find container \"87856f0f10585dd547492ed5711f26f3409ec28f07c808cd58bb845c8f6d7522\": container with ID starting with 87856f0f10585dd547492ed5711f26f3409ec28f07c808cd58bb845c8f6d7522 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.822211 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffd4b07638def8d9e366b9d82e7c3af366fb82361bbee1659a1998e20a8e5b84\": container with ID starting with ffd4b07638def8d9e366b9d82e7c3af366fb82361bbee1659a1998e20a8e5b84 not found: ID does not exist" containerID="ffd4b07638def8d9e366b9d82e7c3af366fb82361bbee1659a1998e20a8e5b84" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.822236 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="ffd4b07638def8d9e366b9d82e7c3af366fb82361bbee1659a1998e20a8e5b84" err="rpc error: code = NotFound desc = could not find container \"ffd4b07638def8d9e366b9d82e7c3af366fb82361bbee1659a1998e20a8e5b84\": container with ID starting with ffd4b07638def8d9e366b9d82e7c3af366fb82361bbee1659a1998e20a8e5b84 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.822493 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e35b3bbc117459ece96a932475aecb63a3482c57612aebbe51e956cad2125e9\": container with ID starting with 5e35b3bbc117459ece96a932475aecb63a3482c57612aebbe51e956cad2125e9 not found: ID does not exist" containerID="5e35b3bbc117459ece96a932475aecb63a3482c57612aebbe51e956cad2125e9" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.822519 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="5e35b3bbc117459ece96a932475aecb63a3482c57612aebbe51e956cad2125e9" err="rpc error: code = NotFound desc = could not find container \"5e35b3bbc117459ece96a932475aecb63a3482c57612aebbe51e956cad2125e9\": container with ID starting with 5e35b3bbc117459ece96a932475aecb63a3482c57612aebbe51e956cad2125e9 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.822751 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af0c9f026ecfe440a5fc88294737d82bee4397499028478469868d8b733da42a\": container with ID starting with af0c9f026ecfe440a5fc88294737d82bee4397499028478469868d8b733da42a not found: ID does not exist" containerID="af0c9f026ecfe440a5fc88294737d82bee4397499028478469868d8b733da42a" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.822777 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="af0c9f026ecfe440a5fc88294737d82bee4397499028478469868d8b733da42a" err="rpc error: code = NotFound desc = could not find container \"af0c9f026ecfe440a5fc88294737d82bee4397499028478469868d8b733da42a\": container with ID starting with af0c9f026ecfe440a5fc88294737d82bee4397499028478469868d8b733da42a not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.823003 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bc32a41872e3fdd7958d7634b2f0a07b2c149a1e7107267df034099986b45b2\": container with ID starting with 0bc32a41872e3fdd7958d7634b2f0a07b2c149a1e7107267df034099986b45b2 not found: ID does not exist" containerID="0bc32a41872e3fdd7958d7634b2f0a07b2c149a1e7107267df034099986b45b2" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.823029 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="0bc32a41872e3fdd7958d7634b2f0a07b2c149a1e7107267df034099986b45b2" err="rpc error: code = NotFound desc = could not find container \"0bc32a41872e3fdd7958d7634b2f0a07b2c149a1e7107267df034099986b45b2\": container with ID starting with 0bc32a41872e3fdd7958d7634b2f0a07b2c149a1e7107267df034099986b45b2 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.823313 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"970a12721ad9e883397cc47eca2c43e0043682e646f53f4100c9f77099dd9974\": container with ID starting with 970a12721ad9e883397cc47eca2c43e0043682e646f53f4100c9f77099dd9974 not found: ID does not exist" containerID="970a12721ad9e883397cc47eca2c43e0043682e646f53f4100c9f77099dd9974" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.823339 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="970a12721ad9e883397cc47eca2c43e0043682e646f53f4100c9f77099dd9974" err="rpc error: code = NotFound desc = could not find container \"970a12721ad9e883397cc47eca2c43e0043682e646f53f4100c9f77099dd9974\": container with ID starting with 970a12721ad9e883397cc47eca2c43e0043682e646f53f4100c9f77099dd9974 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.823578 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46be80933179162fd6984139645807e25e844c3795d674a8b04f734457efcc64\": container with ID starting with 46be80933179162fd6984139645807e25e844c3795d674a8b04f734457efcc64 not found: ID does not exist" containerID="46be80933179162fd6984139645807e25e844c3795d674a8b04f734457efcc64" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.823608 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="46be80933179162fd6984139645807e25e844c3795d674a8b04f734457efcc64" err="rpc error: code = NotFound desc = could not find container \"46be80933179162fd6984139645807e25e844c3795d674a8b04f734457efcc64\": container with ID starting with 46be80933179162fd6984139645807e25e844c3795d674a8b04f734457efcc64 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.823845 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"876bc02802326ac406bee168382dda6d347e8d7eb8b3e236476da5e28e853f80\": container with ID starting with 876bc02802326ac406bee168382dda6d347e8d7eb8b3e236476da5e28e853f80 not found: ID does not exist" containerID="876bc02802326ac406bee168382dda6d347e8d7eb8b3e236476da5e28e853f80" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.823872 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="876bc02802326ac406bee168382dda6d347e8d7eb8b3e236476da5e28e853f80" err="rpc error: code = NotFound desc = could not find container \"876bc02802326ac406bee168382dda6d347e8d7eb8b3e236476da5e28e853f80\": container with ID starting with 876bc02802326ac406bee168382dda6d347e8d7eb8b3e236476da5e28e853f80 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.824101 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7de1317b2bf9c95059be217ab261be612eec0edba15cda55d2880d2366374632\": container with ID starting with 7de1317b2bf9c95059be217ab261be612eec0edba15cda55d2880d2366374632 not found: ID does not exist" containerID="7de1317b2bf9c95059be217ab261be612eec0edba15cda55d2880d2366374632" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.824126 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="7de1317b2bf9c95059be217ab261be612eec0edba15cda55d2880d2366374632" err="rpc error: code = NotFound desc = could not find container \"7de1317b2bf9c95059be217ab261be612eec0edba15cda55d2880d2366374632\": container with ID starting with 7de1317b2bf9c95059be217ab261be612eec0edba15cda55d2880d2366374632 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.824409 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bab9c13d5ff60b04f64724b0be19b71d020c39b00bbce0a42b84b6baf9cf1858\": container with ID starting with bab9c13d5ff60b04f64724b0be19b71d020c39b00bbce0a42b84b6baf9cf1858 not found: ID does not exist" containerID="bab9c13d5ff60b04f64724b0be19b71d020c39b00bbce0a42b84b6baf9cf1858" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.824438 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="bab9c13d5ff60b04f64724b0be19b71d020c39b00bbce0a42b84b6baf9cf1858" err="rpc error: code = NotFound desc = could not find container \"bab9c13d5ff60b04f64724b0be19b71d020c39b00bbce0a42b84b6baf9cf1858\": container with ID starting with bab9c13d5ff60b04f64724b0be19b71d020c39b00bbce0a42b84b6baf9cf1858 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.824681 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce5c1f9ab4c7e7c8e14b69367c484d06af071c61ef13ea7733af64251d5a919e\": container with ID starting with ce5c1f9ab4c7e7c8e14b69367c484d06af071c61ef13ea7733af64251d5a919e not found: ID does not exist" containerID="ce5c1f9ab4c7e7c8e14b69367c484d06af071c61ef13ea7733af64251d5a919e" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.824709 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="ce5c1f9ab4c7e7c8e14b69367c484d06af071c61ef13ea7733af64251d5a919e" err="rpc error: code = NotFound desc = could not find container \"ce5c1f9ab4c7e7c8e14b69367c484d06af071c61ef13ea7733af64251d5a919e\": container with ID starting with ce5c1f9ab4c7e7c8e14b69367c484d06af071c61ef13ea7733af64251d5a919e not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.824957 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff00bd198c611d428013e67929b1264e892e200f10b113006fa289c79f2c6c50\": container with ID starting with ff00bd198c611d428013e67929b1264e892e200f10b113006fa289c79f2c6c50 not found: ID does not exist" containerID="ff00bd198c611d428013e67929b1264e892e200f10b113006fa289c79f2c6c50" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.824982 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="ff00bd198c611d428013e67929b1264e892e200f10b113006fa289c79f2c6c50" err="rpc error: code = NotFound desc = could not find container \"ff00bd198c611d428013e67929b1264e892e200f10b113006fa289c79f2c6c50\": container with ID starting with ff00bd198c611d428013e67929b1264e892e200f10b113006fa289c79f2c6c50 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.825245 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e9aa74411f0cbe12125f1767253aa311244845770097efc2342f5f2213a5f99\": container with ID starting with 0e9aa74411f0cbe12125f1767253aa311244845770097efc2342f5f2213a5f99 not found: ID does not exist" containerID="0e9aa74411f0cbe12125f1767253aa311244845770097efc2342f5f2213a5f99" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.825283 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="0e9aa74411f0cbe12125f1767253aa311244845770097efc2342f5f2213a5f99" err="rpc error: code = NotFound desc = could not find container \"0e9aa74411f0cbe12125f1767253aa311244845770097efc2342f5f2213a5f99\": container with ID starting with 0e9aa74411f0cbe12125f1767253aa311244845770097efc2342f5f2213a5f99 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.825600 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f81d81f334991a1a28d4bd7f1196fbd098494ebccd408af1173ec1fe9616bd86\": container with ID starting with f81d81f334991a1a28d4bd7f1196fbd098494ebccd408af1173ec1fe9616bd86 not found: ID does not exist" containerID="f81d81f334991a1a28d4bd7f1196fbd098494ebccd408af1173ec1fe9616bd86" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.825627 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="f81d81f334991a1a28d4bd7f1196fbd098494ebccd408af1173ec1fe9616bd86" err="rpc error: code = NotFound desc = could not find container \"f81d81f334991a1a28d4bd7f1196fbd098494ebccd408af1173ec1fe9616bd86\": container with ID starting with f81d81f334991a1a28d4bd7f1196fbd098494ebccd408af1173ec1fe9616bd86 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.825880 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d7597386b38676c039e5b6aac3f7c678b4af05333bf527578cbc167c1f3ec5c\": container with ID starting with 1d7597386b38676c039e5b6aac3f7c678b4af05333bf527578cbc167c1f3ec5c not found: ID does not exist" containerID="1d7597386b38676c039e5b6aac3f7c678b4af05333bf527578cbc167c1f3ec5c" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.825906 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="1d7597386b38676c039e5b6aac3f7c678b4af05333bf527578cbc167c1f3ec5c" err="rpc error: code = NotFound desc = could not find container \"1d7597386b38676c039e5b6aac3f7c678b4af05333bf527578cbc167c1f3ec5c\": container with ID starting with 1d7597386b38676c039e5b6aac3f7c678b4af05333bf527578cbc167c1f3ec5c not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.826168 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81d748520620965f53fabe92f19a5bb3c5d4e2f876e096dcb82692d2f2f8f610\": container with ID starting with 81d748520620965f53fabe92f19a5bb3c5d4e2f876e096dcb82692d2f2f8f610 not found: ID does not exist" containerID="81d748520620965f53fabe92f19a5bb3c5d4e2f876e096dcb82692d2f2f8f610" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.826200 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="81d748520620965f53fabe92f19a5bb3c5d4e2f876e096dcb82692d2f2f8f610" err="rpc error: code = NotFound desc = could not find container \"81d748520620965f53fabe92f19a5bb3c5d4e2f876e096dcb82692d2f2f8f610\": container with ID starting with 81d748520620965f53fabe92f19a5bb3c5d4e2f876e096dcb82692d2f2f8f610 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.826495 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b77a73a551754a882d136a50cbe7d425606b96a3c681a87e9ec959b8bfe08ce\": container with ID starting with 3b77a73a551754a882d136a50cbe7d425606b96a3c681a87e9ec959b8bfe08ce not found: ID does not exist" containerID="3b77a73a551754a882d136a50cbe7d425606b96a3c681a87e9ec959b8bfe08ce" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.826524 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="3b77a73a551754a882d136a50cbe7d425606b96a3c681a87e9ec959b8bfe08ce" err="rpc error: code = NotFound desc = could not find container \"3b77a73a551754a882d136a50cbe7d425606b96a3c681a87e9ec959b8bfe08ce\": container with ID starting with 3b77a73a551754a882d136a50cbe7d425606b96a3c681a87e9ec959b8bfe08ce not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.826763 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c8fab5cf2da97490127bff84313a78ffc021feac020252239d18a75f21cbc6b\": container with ID starting with 5c8fab5cf2da97490127bff84313a78ffc021feac020252239d18a75f21cbc6b not found: ID does not exist" containerID="5c8fab5cf2da97490127bff84313a78ffc021feac020252239d18a75f21cbc6b" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.826789 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="5c8fab5cf2da97490127bff84313a78ffc021feac020252239d18a75f21cbc6b" err="rpc error: code = NotFound desc = could not find container \"5c8fab5cf2da97490127bff84313a78ffc021feac020252239d18a75f21cbc6b\": container with ID starting with 5c8fab5cf2da97490127bff84313a78ffc021feac020252239d18a75f21cbc6b not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.827037 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14af629856736bb66f0a281297394a8dac31178dfcdb2788d39f3744b283ba3e\": container with ID starting with 14af629856736bb66f0a281297394a8dac31178dfcdb2788d39f3744b283ba3e not found: ID does not exist" containerID="14af629856736bb66f0a281297394a8dac31178dfcdb2788d39f3744b283ba3e" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.827064 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="14af629856736bb66f0a281297394a8dac31178dfcdb2788d39f3744b283ba3e" err="rpc error: code = NotFound desc = could not find container \"14af629856736bb66f0a281297394a8dac31178dfcdb2788d39f3744b283ba3e\": container with ID starting with 14af629856736bb66f0a281297394a8dac31178dfcdb2788d39f3744b283ba3e not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.827338 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a04d63e53a7fcdfd5ab131841815a2f2677c7f07a565fb1c56e7fe2739ce6414\": container with ID starting with a04d63e53a7fcdfd5ab131841815a2f2677c7f07a565fb1c56e7fe2739ce6414 not found: ID does not exist" containerID="a04d63e53a7fcdfd5ab131841815a2f2677c7f07a565fb1c56e7fe2739ce6414" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.827366 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="a04d63e53a7fcdfd5ab131841815a2f2677c7f07a565fb1c56e7fe2739ce6414" err="rpc error: code = NotFound desc = could not find container \"a04d63e53a7fcdfd5ab131841815a2f2677c7f07a565fb1c56e7fe2739ce6414\": container with ID starting with a04d63e53a7fcdfd5ab131841815a2f2677c7f07a565fb1c56e7fe2739ce6414 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.827603 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"183543c83fed2e18744fbae52d6331fb21aa60a9b2756ab8cb9cfbc9fdfbad3f\": container with ID starting with 183543c83fed2e18744fbae52d6331fb21aa60a9b2756ab8cb9cfbc9fdfbad3f not found: ID does not exist" containerID="183543c83fed2e18744fbae52d6331fb21aa60a9b2756ab8cb9cfbc9fdfbad3f" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.827629 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="183543c83fed2e18744fbae52d6331fb21aa60a9b2756ab8cb9cfbc9fdfbad3f" err="rpc error: code = NotFound desc = could not find container \"183543c83fed2e18744fbae52d6331fb21aa60a9b2756ab8cb9cfbc9fdfbad3f\": container with ID starting with 183543c83fed2e18744fbae52d6331fb21aa60a9b2756ab8cb9cfbc9fdfbad3f not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.827869 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"474834dd0b54abe011163e7458744c9ca3a52b894e6a07cfb86b009916a74367\": container with ID starting with 474834dd0b54abe011163e7458744c9ca3a52b894e6a07cfb86b009916a74367 not found: ID does not exist" containerID="474834dd0b54abe011163e7458744c9ca3a52b894e6a07cfb86b009916a74367" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.827895 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="474834dd0b54abe011163e7458744c9ca3a52b894e6a07cfb86b009916a74367" err="rpc error: code = NotFound desc = could not find container \"474834dd0b54abe011163e7458744c9ca3a52b894e6a07cfb86b009916a74367\": container with ID starting with 474834dd0b54abe011163e7458744c9ca3a52b894e6a07cfb86b009916a74367 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.828155 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ba3e4783da8b547347f29221e83019680bad22ec9f3f891d05c7beed9fcbcad\": container with ID starting with 0ba3e4783da8b547347f29221e83019680bad22ec9f3f891d05c7beed9fcbcad not found: ID does not exist" containerID="0ba3e4783da8b547347f29221e83019680bad22ec9f3f891d05c7beed9fcbcad" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.828183 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="0ba3e4783da8b547347f29221e83019680bad22ec9f3f891d05c7beed9fcbcad" err="rpc error: code = NotFound desc = could not find container \"0ba3e4783da8b547347f29221e83019680bad22ec9f3f891d05c7beed9fcbcad\": container with ID starting with 0ba3e4783da8b547347f29221e83019680bad22ec9f3f891d05c7beed9fcbcad not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.828447 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"668fca10db53532d614671680bb989d57b924efcbca09e6ef2b7e4895ccd7abb\": container with ID starting with 668fca10db53532d614671680bb989d57b924efcbca09e6ef2b7e4895ccd7abb not found: ID does not exist" containerID="668fca10db53532d614671680bb989d57b924efcbca09e6ef2b7e4895ccd7abb" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.828476 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="668fca10db53532d614671680bb989d57b924efcbca09e6ef2b7e4895ccd7abb" err="rpc error: code = NotFound desc = could not find container \"668fca10db53532d614671680bb989d57b924efcbca09e6ef2b7e4895ccd7abb\": container with ID starting with 668fca10db53532d614671680bb989d57b924efcbca09e6ef2b7e4895ccd7abb not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.828745 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"722bd8448df2ddebd63be975ab1961567271a1f1cca344e95f989bcc07d89d6e\": container with ID starting with 722bd8448df2ddebd63be975ab1961567271a1f1cca344e95f989bcc07d89d6e not found: ID does not exist" containerID="722bd8448df2ddebd63be975ab1961567271a1f1cca344e95f989bcc07d89d6e" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.828776 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="722bd8448df2ddebd63be975ab1961567271a1f1cca344e95f989bcc07d89d6e" err="rpc error: code = NotFound desc = could not find container \"722bd8448df2ddebd63be975ab1961567271a1f1cca344e95f989bcc07d89d6e\": container with ID starting with 722bd8448df2ddebd63be975ab1961567271a1f1cca344e95f989bcc07d89d6e not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.829019 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4d6579886d4371779b80ec613338f22c6a967a0e93854fba505389a0fc61b57\": container with ID starting with f4d6579886d4371779b80ec613338f22c6a967a0e93854fba505389a0fc61b57 not found: ID does not exist" containerID="f4d6579886d4371779b80ec613338f22c6a967a0e93854fba505389a0fc61b57" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.829053 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="f4d6579886d4371779b80ec613338f22c6a967a0e93854fba505389a0fc61b57" err="rpc error: code = NotFound desc = could not find container \"f4d6579886d4371779b80ec613338f22c6a967a0e93854fba505389a0fc61b57\": container with ID starting with f4d6579886d4371779b80ec613338f22c6a967a0e93854fba505389a0fc61b57 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.829341 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"205169b62d7080fa5dcfe7076f14dc28577ed4c769ffe43d35fb73df48f1cd2e\": container with ID starting with 205169b62d7080fa5dcfe7076f14dc28577ed4c769ffe43d35fb73df48f1cd2e not found: ID does not exist" containerID="205169b62d7080fa5dcfe7076f14dc28577ed4c769ffe43d35fb73df48f1cd2e" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.829370 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="205169b62d7080fa5dcfe7076f14dc28577ed4c769ffe43d35fb73df48f1cd2e" err="rpc error: code = NotFound desc = could not find container \"205169b62d7080fa5dcfe7076f14dc28577ed4c769ffe43d35fb73df48f1cd2e\": container with ID starting with 205169b62d7080fa5dcfe7076f14dc28577ed4c769ffe43d35fb73df48f1cd2e not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.829619 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d14df105d1078625422b2a59ff6a8bdadf9cd80f1516885c6074d5a7d6d78616\": container with ID starting with d14df105d1078625422b2a59ff6a8bdadf9cd80f1516885c6074d5a7d6d78616 not found: ID does not exist" containerID="d14df105d1078625422b2a59ff6a8bdadf9cd80f1516885c6074d5a7d6d78616" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.829652 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="d14df105d1078625422b2a59ff6a8bdadf9cd80f1516885c6074d5a7d6d78616" err="rpc error: code = NotFound desc = could not find container \"d14df105d1078625422b2a59ff6a8bdadf9cd80f1516885c6074d5a7d6d78616\": container with ID starting with d14df105d1078625422b2a59ff6a8bdadf9cd80f1516885c6074d5a7d6d78616 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.829888 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00f696cbb09d5836744a030a84da636141f60f03f340f69b27a9677ead91bf17\": container with ID starting with 00f696cbb09d5836744a030a84da636141f60f03f340f69b27a9677ead91bf17 not found: ID does not exist" containerID="00f696cbb09d5836744a030a84da636141f60f03f340f69b27a9677ead91bf17" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.829914 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="00f696cbb09d5836744a030a84da636141f60f03f340f69b27a9677ead91bf17" err="rpc error: code = NotFound desc = could not find container \"00f696cbb09d5836744a030a84da636141f60f03f340f69b27a9677ead91bf17\": container with ID starting with 00f696cbb09d5836744a030a84da636141f60f03f340f69b27a9677ead91bf17 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.830165 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b60f1d912768234f658c95fc37c5638dd28e2bbc01c64bb2abb830967916a74\": container with ID starting with 5b60f1d912768234f658c95fc37c5638dd28e2bbc01c64bb2abb830967916a74 not found: ID does not exist" containerID="5b60f1d912768234f658c95fc37c5638dd28e2bbc01c64bb2abb830967916a74" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.830192 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="5b60f1d912768234f658c95fc37c5638dd28e2bbc01c64bb2abb830967916a74" err="rpc error: code = NotFound desc = could not find container \"5b60f1d912768234f658c95fc37c5638dd28e2bbc01c64bb2abb830967916a74\": container with ID starting with 5b60f1d912768234f658c95fc37c5638dd28e2bbc01c64bb2abb830967916a74 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.830450 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e5bb1b82e50570e1238f090a0c69c8ce8c4924181c9fe1d3c2e79bbcaf04cf7\": container with ID starting with 0e5bb1b82e50570e1238f090a0c69c8ce8c4924181c9fe1d3c2e79bbcaf04cf7 not found: ID does not exist" containerID="0e5bb1b82e50570e1238f090a0c69c8ce8c4924181c9fe1d3c2e79bbcaf04cf7" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.830479 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="0e5bb1b82e50570e1238f090a0c69c8ce8c4924181c9fe1d3c2e79bbcaf04cf7" err="rpc error: code = NotFound desc = could not find container \"0e5bb1b82e50570e1238f090a0c69c8ce8c4924181c9fe1d3c2e79bbcaf04cf7\": container with ID starting with 0e5bb1b82e50570e1238f090a0c69c8ce8c4924181c9fe1d3c2e79bbcaf04cf7 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.830716 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba948b0e959411da071a39a920e3e612512d962feea740e6d05b34f3da0f9919\": container with ID starting with ba948b0e959411da071a39a920e3e612512d962feea740e6d05b34f3da0f9919 not found: ID does not exist" containerID="ba948b0e959411da071a39a920e3e612512d962feea740e6d05b34f3da0f9919" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.830741 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="ba948b0e959411da071a39a920e3e612512d962feea740e6d05b34f3da0f9919" err="rpc error: code = NotFound desc = could not find container \"ba948b0e959411da071a39a920e3e612512d962feea740e6d05b34f3da0f9919\": container with ID starting with ba948b0e959411da071a39a920e3e612512d962feea740e6d05b34f3da0f9919 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.830980 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1806ad111afb95a724d10bd8c46e3bd826a078fefa581add2dbef820b5ef1cb4\": container with ID starting with 1806ad111afb95a724d10bd8c46e3bd826a078fefa581add2dbef820b5ef1cb4 not found: ID does not exist" containerID="1806ad111afb95a724d10bd8c46e3bd826a078fefa581add2dbef820b5ef1cb4" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.831006 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="1806ad111afb95a724d10bd8c46e3bd826a078fefa581add2dbef820b5ef1cb4" err="rpc error: code = NotFound desc = could not find container \"1806ad111afb95a724d10bd8c46e3bd826a078fefa581add2dbef820b5ef1cb4\": container with ID starting with 1806ad111afb95a724d10bd8c46e3bd826a078fefa581add2dbef820b5ef1cb4 not found: ID does not exist" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: E0105 09:45:40.831261 2588 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d221c9b7969a4393326f6db0244fa128281bf5eb98e6da040c6d3d1135a666d\": container with ID starting with 3d221c9b7969a4393326f6db0244fa128281bf5eb98e6da040c6d3d1135a666d not found: ID does not exist" containerID="3d221c9b7969a4393326f6db0244fa128281bf5eb98e6da040c6d3d1135a666d" Jan 05 09:45:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:40.831296 2588 kuberuntime_gc.go:361] "Error getting ContainerStatus for containerID" containerID="3d221c9b7969a4393326f6db0244fa128281bf5eb98e6da040c6d3d1135a666d" err="rpc error: code = NotFound desc = could not find container \"3d221c9b7969a4393326f6db0244fa128281bf5eb98e6da040c6d3d1135a666d\": container with ID starting with 3d221c9b7969a4393326f6db0244fa128281bf5eb98e6da040c6d3d1135a666d not found: ID does not exist" Jan 05 09:45:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:42.011198 2588 dynamic_cafile_content.go:211] "Failed to remove file watch, it may have been deleted" file="/etc/kubernetes/kubelet-ca.crt" err="can't remove non-existent inotify watch for: /etc/kubernetes/kubelet-ca.crt" Jan 05 09:45:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:44.115880 2588 plugin_watcher.go:215] "Removing socket path from desired state cache" path="/var/lib/kubelet/plugins_registry/diskplugin.csi.alibabacloud.com-reg.sock" Jan 05 09:45:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[2588]: I0105 09:45:44.164077 2588 dynamic_cafile_content.go:171] "Shutting down controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 05 09:45:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Stopping Kubernetes Kubelet... Jan 05 09:45:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: kubelet.service: Succeeded. Jan 05 09:45:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Stopped Kubernetes Kubelet. Jan 05 09:45:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: kubelet.service: Consumed 16.553s CPU time Jan 05 09:46:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:46:02.552595684Z" level=error msg="Failed to cleanup (probably retrying): failed to destroy network for pod sandbox k8s_prometheus-k8s-0_openshift-monitoring_4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879_0(04ef5a2380724b396d64799983d162496bd934c4c1c29b3813491756b43870a0): error removing pod openshift-monitoring_prometheus-k8s-0 from CNI network \"multus-cni-network\": plugin type=\"multus\" name=\"multus-cni-network\" failed (delete): Multus: [openshift-monitoring/prometheus-k8s-0/4ad4cdcf-3061-4f41-9ab3-5f7f06e2b879]: PollImmediate error waiting for ReadinessIndicatorFile (on del): timed out waiting for the condition" Jan 05 09:46:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Stopping Container Runtime Interface for OCI (CRI-O)... Jan 05 09:46:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[2545]: time="2023-01-05 09:46:04.203111213Z" level=error msg="Failed to update container state for aee14ddf06d77378c3afc5489f1ac566eb37f4e3542ad949a8b18afeff164f94: `/usr/bin/runc --root /run/runc --systemd-cgroup state aee14ddf06d77378c3afc5489f1ac566eb37f4e3542ad949a8b18afeff164f94` failed: : signal: terminated" Jan 05 09:46:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: crio.service: Succeeded. Jan 05 09:46:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Stopped Container Runtime Interface for OCI (CRI-O). Jan 05 09:46:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: crio.service: Consumed 16.007s CPU time -- Reboot -- Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.243458994Z" level=info msg="Starting CRI-O, version: 1.25.1-5.rhaos4.12.git6005903.el8, git: unknown(clean)" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.243684577Z" level=info msg="Node configuration value for hugetlb cgroup is true" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.243698299Z" level=info msg="Node configuration value for pid cgroup is true" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.243757089Z" level=info msg="Node configuration value for memoryswap cgroup is true" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.243767050Z" level=info msg="Node configuration value for cgroup v2 is false" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.257371824Z" level=info msg="Node configuration value for systemd CollectMode is true" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.265863386Z" level=info msg="Node configuration value for systemd AllowedCPUs is true" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.271844988Z" level=info msg="Using default capabilities: CAP_CHOWN, CAP_DAC_OVERRIDE, CAP_FSETID, CAP_FOWNER, CAP_SETGID, CAP_SETUID, CAP_SETPCAP, CAP_NET_BIND_SERVICE, CAP_KILL" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Starting Container Runtime Interface for OCI (CRI-O)... Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.349984212Z" level=info msg="Checkpoint/restore support disabled" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.350146132Z" level=info msg="Using seccomp default profile when unspecified: true" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.350207074Z" level=info msg="Using the internal default seccomp profile" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.350264527Z" level=info msg="AppArmor is disabled by the system or at CRI-O build-time" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.350337009Z" level=info msg="No blockio config file specified, blockio not configured" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.350398897Z" level=info msg="RDT not available in the host system" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.356010987Z" level=info msg="Conmon does support the --sync option" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.356030252Z" level=info msg="Conmon does support the --log-global-size-max option" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.359663505Z" level=info msg="Conmon does support the --sync option" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.359679536Z" level=info msg="Conmon does support the --log-global-size-max option" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.470009462Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.470036923Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.521822988Z" level=warning msg="Could not restore sandbox 93662ac8682d21720eb888cc865f999977edd4dcd1e585e0bae6c3aeeb2d9bba: failed to Statfs \"/var/run/netns/dc235f8e-171c-4950-87de-fccc48b4525b\": no such file or directory" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.535492215Z" level=warning msg="Deleting all containers under sandbox 93662ac8682d21720eb888cc865f999977edd4dcd1e585e0bae6c3aeeb2d9bba since it could not be restored" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.542230838Z" level=warning msg="Could not restore sandbox 2f42fdd8caa6c772e33c09307d855f2c6b126d16a962580bfad9af47fa893ef2: failed to Statfs \"/var/run/netns/db10749b-0af7-4faa-bbc1-64769e589a18\": no such file or directory" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.551182728Z" level=warning msg="Deleting all containers under sandbox 2f42fdd8caa6c772e33c09307d855f2c6b126d16a962580bfad9af47fa893ef2 since it could not be restored" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.559805816Z" level=warning msg="Could not restore sandbox 06a6d838dd853ce187ccf364676fd4e36976cae02f20a98aeda4b59df6df2d93: failed to Statfs \"/var/run/netns/66584c74-3d2e-45f9-91c5-50bf436ccb57\": no such file or directory" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.567921889Z" level=warning msg="Deleting all containers under sandbox 06a6d838dd853ce187ccf364676fd4e36976cae02f20a98aeda4b59df6df2d93 since it could not be restored" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.591111951Z" level=warning msg="Could not restore sandbox acd08c77f0076092e020bedcc20e2c94faab21df9b7d5516eba6f0dbe370935b: failed to Statfs \"/var/run/netns/173db4d2-8fc3-4eb4-8fed-72463bf97e83\": no such file or directory" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.600570291Z" level=warning msg="Deleting all containers under sandbox acd08c77f0076092e020bedcc20e2c94faab21df9b7d5516eba6f0dbe370935b since it could not be restored" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.624481754Z" level=warning msg="Could not restore sandbox ae18be6d794af60a14ce35ff3392698b0b7e426a5128047f31455ade1f2b79dd: failed to Statfs \"/var/run/netns/893e3f97-f26d-47c1-9aed-64615483b236\": no such file or directory" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.631340857Z" level=warning msg="Deleting all containers under sandbox ae18be6d794af60a14ce35ff3392698b0b7e426a5128047f31455ade1f2b79dd since it could not be restored" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.647032712Z" level=warning msg="Could not restore sandbox bb5b6562ae64906e3450c53a90bf4df6ea55af8ce39868944ba2cdcf7eab1188: failed to Statfs \"/var/run/netns/196e4ac5-eb4d-4b9b-a870-07aff8b9b61e\": no such file or directory" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.653942855Z" level=warning msg="Deleting all containers under sandbox bb5b6562ae64906e3450c53a90bf4df6ea55af8ce39868944ba2cdcf7eab1188 since it could not be restored" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.662933220Z" level=warning msg="Could not restore sandbox e0df29cbf3197efe5e4c221145d0f72e155c977f3d50c18c2f38862f24a5522e: failed to Statfs \"/var/run/netns/6b604cc9-ab4b-4fe2-9286-1f452782c20d\": no such file or directory" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.668848207Z" level=warning msg="Deleting all containers under sandbox e0df29cbf3197efe5e4c221145d0f72e155c977f3d50c18c2f38862f24a5522e since it could not be restored" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.676571802Z" level=warning msg="Could not restore sandbox 16cbcad68c73abf56080eb14225620b7f239901aad0b8916effb3d6a9dc1a771: failed to Statfs \"/var/run/netns/1b0a9a52-5775-4a46-b12e-300f9faa4092\": no such file or directory" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.683758901Z" level=warning msg="Deleting all containers under sandbox 16cbcad68c73abf56080eb14225620b7f239901aad0b8916effb3d6a9dc1a771 since it could not be restored" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.703774659Z" level=warning msg="Could not restore sandbox 4f0a921bd0d3c8f309751665c310e818034ecd3252406f2a101ca811d04377c5: failed to Statfs \"/var/run/netns/7b20456c-f122-4dc5-ba26-66ea4aeb44a5\": no such file or directory" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.710943126Z" level=warning msg="Deleting all containers under sandbox 4f0a921bd0d3c8f309751665c310e818034ecd3252406f2a101ca811d04377c5 since it could not be restored" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.722802373Z" level=warning msg="Could not restore sandbox e9053289b0b05a28538d40135f5851bcbb4fef2e500409bd4a7c787c89dc9515: failed to Statfs \"/var/run/netns/75290ffc-f543-4fcc-80ac-c0a9794ef66f\": no such file or directory" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.728916250Z" level=warning msg="Deleting all containers under sandbox e9053289b0b05a28538d40135f5851bcbb4fef2e500409bd4a7c787c89dc9515 since it could not be restored" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.769640507Z" level=warning msg="Could not restore sandbox 91eb7fe6d0c596f220833b072fadd2715083a9931fdb3ed2bbf7986380dce490: failed to Statfs \"/var/run/netns/a97e603f-5d7c-4b89-8780-e83788b64081\": no such file or directory" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.774679575Z" level=warning msg="Deleting all containers under sandbox 91eb7fe6d0c596f220833b072fadd2715083a9931fdb3ed2bbf7986380dce490 since it could not be restored" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.780129473Z" level=warning msg="Could not restore sandbox 44ef9ac88935fc68d10995ab1204d3f92afbedeae9fabaf8c0abb737181fa9fd: failed to Statfs \"/var/run/netns/5a7d8972-1e35-45cd-add2-715bbc661172\": no such file or directory" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.785285429Z" level=warning msg="Deleting all containers under sandbox 44ef9ac88935fc68d10995ab1204d3f92afbedeae9fabaf8c0abb737181fa9fd since it could not be restored" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.796479257Z" level=warning msg="Could not restore sandbox df03e4f874ec90019c812fc416c2d12340d64e387a2a542bdc3beee879bc31bd: failed to Statfs \"/var/run/netns/13c36503-2dc5-4ab8-a8bd-dd131abd1770\": no such file or directory" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.803279398Z" level=warning msg="Deleting all containers under sandbox df03e4f874ec90019c812fc416c2d12340d64e387a2a542bdc3beee879bc31bd since it could not be restored" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.808445603Z" level=warning msg="Error encountered when checking whether cri-o should wipe containers: open /var/run/crio/version: no such file or directory" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.808562436Z" level=info msg="cleanup sandbox network" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.809184834Z" level=info msg="Successfully cleaned up network for pod e9053289b0b05a28538d40135f5851bcbb4fef2e500409bd4a7c787c89dc9515" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.809200655Z" level=info msg="cleanup sandbox network" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.811524707Z" level=info msg="Serving metrics on :9537 via HTTP" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Started Container Runtime Interface for OCI (CRI-O). Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.812511132Z" level=info msg="Got pod network &{Name:dns-default-2pfzf Namespace:openshift-dns ID:4f0a921bd0d3c8f309751665c310e818034ecd3252406f2a101ca811d04377c5 UID:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:35.813058655Z" level=info msg="Deleting pod openshift-dns_dns-default-2pfzf from CNI network \"multus-cni-network\" (type=multus)" Jan 05 17:46:35 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Starting Kubernetes Kubelet... Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: Flag --container-runtime has been deprecated, will be removed in 1.27 as the only valid value is 'remote' Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: Flag --cloud-provider has been deprecated, will be removed in 1.25 or later, in favor of removing cloud provider code from Kubelet. Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: Flag --provider-id has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.27. Image garbage collector will get sandbox image information from CRI. Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.201711 1977 server.go:200] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: Flag --container-runtime has been deprecated, will be removed in 1.27 as the only valid value is 'remote' Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: Flag --cloud-provider has been deprecated, will be removed in 1.25 or later, in favor of removing cloud provider code from Kubelet. Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: Flag --provider-id has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.27. Image garbage collector will get sandbox image information from CRI. Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203807 1977 flags.go:64] FLAG: --add-dir-header="false" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203819 1977 flags.go:64] FLAG: --address="0.0.0.0" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203824 1977 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203834 1977 flags.go:64] FLAG: --alsologtostderr="false" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203837 1977 flags.go:64] FLAG: --anonymous-auth="true" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203841 1977 flags.go:64] FLAG: --application-metrics-count-limit="100" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203845 1977 flags.go:64] FLAG: --authentication-token-webhook="false" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203852 1977 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203857 1977 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203862 1977 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203867 1977 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203873 1977 flags.go:64] FLAG: --azure-container-registry-config="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203878 1977 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203883 1977 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203887 1977 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203890 1977 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203894 1977 flags.go:64] FLAG: --cgroup-root="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203898 1977 flags.go:64] FLAG: --cgroups-per-qos="true" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203901 1977 flags.go:64] FLAG: --client-ca-file="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203904 1977 flags.go:64] FLAG: --cloud-config="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203907 1977 flags.go:64] FLAG: --cloud-provider="external" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203911 1977 flags.go:64] FLAG: --cluster-dns="[]" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203918 1977 flags.go:64] FLAG: --cluster-domain="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203921 1977 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203925 1977 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203929 1977 flags.go:64] FLAG: --container-log-max-files="5" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203935 1977 flags.go:64] FLAG: --container-log-max-size="10Mi" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203941 1977 flags.go:64] FLAG: --container-runtime="remote" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203945 1977 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203949 1977 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203952 1977 flags.go:64] FLAG: --containerd-namespace="k8s.io" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203956 1977 flags.go:64] FLAG: --contention-profiling="false" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203959 1977 flags.go:64] FLAG: --cpu-cfs-quota="true" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203963 1977 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203967 1977 flags.go:64] FLAG: --cpu-manager-policy="none" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203970 1977 flags.go:64] FLAG: --cpu-manager-policy-options="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203975 1977 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203978 1977 flags.go:64] FLAG: --enable-controller-attach-detach="true" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203981 1977 flags.go:64] FLAG: --enable-debugging-handlers="true" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203984 1977 flags.go:64] FLAG: --enable-load-reader="false" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203988 1977 flags.go:64] FLAG: --enable-server="true" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.203992 1977 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204004 1977 flags.go:64] FLAG: --event-burst="10" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204008 1977 flags.go:64] FLAG: --event-qps="5" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204011 1977 flags.go:64] FLAG: --event-storage-age-limit="default=0" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204014 1977 flags.go:64] FLAG: --event-storage-event-limit="default=0" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204021 1977 flags.go:64] FLAG: --eviction-hard="imagefs.available<15%,memory.available<100Mi,nodefs.available<10%,nodefs.inodesFree<5%" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204031 1977 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204034 1977 flags.go:64] FLAG: --eviction-minimum-reclaim="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204037 1977 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204041 1977 flags.go:64] FLAG: --eviction-soft="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204044 1977 flags.go:64] FLAG: --eviction-soft-grace-period="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204047 1977 flags.go:64] FLAG: --exit-on-lock-contention="false" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204051 1977 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204054 1977 flags.go:64] FLAG: --experimental-mounter-path="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204057 1977 flags.go:64] FLAG: --fail-swap-on="true" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204061 1977 flags.go:64] FLAG: --feature-gates="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204065 1977 flags.go:64] FLAG: --file-check-frequency="20s" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204070 1977 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204073 1977 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204077 1977 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204080 1977 flags.go:64] FLAG: --healthz-port="10248" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204084 1977 flags.go:64] FLAG: --help="false" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204090 1977 flags.go:64] FLAG: --hostname-override="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204094 1977 flags.go:64] FLAG: --housekeeping-interval="10s" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204100 1977 flags.go:64] FLAG: --http-check-frequency="20s" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204104 1977 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204108 1977 flags.go:64] FLAG: --image-credential-provider-config="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204111 1977 flags.go:64] FLAG: --image-gc-high-threshold="85" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204114 1977 flags.go:64] FLAG: --image-gc-low-threshold="80" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204117 1977 flags.go:64] FLAG: --image-service-endpoint="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204120 1977 flags.go:64] FLAG: --iptables-drop-bit="15" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204123 1977 flags.go:64] FLAG: --iptables-masquerade-bit="14" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204127 1977 flags.go:64] FLAG: --keep-terminated-pod-volumes="false" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204131 1977 flags.go:64] FLAG: --kernel-memcg-notification="false" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204134 1977 flags.go:64] FLAG: --kube-api-burst="10" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204138 1977 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204141 1977 flags.go:64] FLAG: --kube-api-qps="5" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204144 1977 flags.go:64] FLAG: --kube-reserved="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204148 1977 flags.go:64] FLAG: --kube-reserved-cgroup="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204151 1977 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204154 1977 flags.go:64] FLAG: --kubelet-cgroups="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204157 1977 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204161 1977 flags.go:64] FLAG: --lock-file="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204167 1977 flags.go:64] FLAG: --log-backtrace-at=":0" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204172 1977 flags.go:64] FLAG: --log-cadvisor-usage="false" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204175 1977 flags.go:64] FLAG: --log-dir="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204178 1977 flags.go:64] FLAG: --log-file="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204181 1977 flags.go:64] FLAG: --log-file-max-size="1800" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204185 1977 flags.go:64] FLAG: --log-flush-frequency="5s" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204189 1977 flags.go:64] FLAG: --log-json-info-buffer-size="0" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204201 1977 flags.go:64] FLAG: --log-json-split-stream="false" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204204 1977 flags.go:64] FLAG: --logging-format="text" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204207 1977 flags.go:64] FLAG: --logtostderr="true" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204210 1977 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204215 1977 flags.go:64] FLAG: --make-iptables-util-chains="true" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204218 1977 flags.go:64] FLAG: --manifest-url="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204222 1977 flags.go:64] FLAG: --manifest-url-header="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204231 1977 flags.go:64] FLAG: --master-service-namespace="default" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204235 1977 flags.go:64] FLAG: --max-housekeeping-interval="15s" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204239 1977 flags.go:64] FLAG: --max-open-files="1000000" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204243 1977 flags.go:64] FLAG: --max-pods="110" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204246 1977 flags.go:64] FLAG: --maximum-dead-containers="-1" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204250 1977 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204256 1977 flags.go:64] FLAG: --memory-manager-policy="None" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204260 1977 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204266 1977 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204269 1977 flags.go:64] FLAG: --node-ip="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204272 1977 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/worker=,node.openshift.io/os_id=rhcos" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204278 1977 flags.go:64] FLAG: --node-status-max-images="50" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204282 1977 flags.go:64] FLAG: --node-status-update-frequency="10s" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204285 1977 flags.go:64] FLAG: --one-output="false" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204288 1977 flags.go:64] FLAG: --oom-score-adj="-999" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204292 1977 flags.go:64] FLAG: --pod-cidr="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204295 1977 flags.go:64] FLAG: --pod-infra-container-image="registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204316 1977 flags.go:64] FLAG: --pod-manifest-path="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204322 1977 flags.go:64] FLAG: --pod-max-pids="-1" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204327 1977 flags.go:64] FLAG: --pods-per-core="0" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204332 1977 flags.go:64] FLAG: --port="10250" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204337 1977 flags.go:64] FLAG: --protect-kernel-defaults="false" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204343 1977 flags.go:64] FLAG: --provider-id="alicloud://us-east-1.i-0xif5dui7khr4rm1wp68" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204349 1977 flags.go:64] FLAG: --qos-reserved="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204355 1977 flags.go:64] FLAG: --read-only-port="10255" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204369 1977 flags.go:64] FLAG: --register-node="true" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204375 1977 flags.go:64] FLAG: --register-schedulable="true" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204380 1977 flags.go:64] FLAG: --register-with-taints="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204384 1977 flags.go:64] FLAG: --registry-burst="10" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204388 1977 flags.go:64] FLAG: --registry-qps="5" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204391 1977 flags.go:64] FLAG: --reserved-cpus="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204394 1977 flags.go:64] FLAG: --reserved-memory="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204398 1977 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204402 1977 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204406 1977 flags.go:64] FLAG: --rotate-certificates="false" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204409 1977 flags.go:64] FLAG: --rotate-server-certificates="false" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204412 1977 flags.go:64] FLAG: --runonce="false" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204415 1977 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204419 1977 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204423 1977 flags.go:64] FLAG: --seccomp-default="false" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204426 1977 flags.go:64] FLAG: --serialize-image-pulls="true" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204431 1977 flags.go:64] FLAG: --skip-headers="false" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204435 1977 flags.go:64] FLAG: --skip-log-headers="false" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204438 1977 flags.go:64] FLAG: --stderrthreshold="2" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204441 1977 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204448 1977 flags.go:64] FLAG: --storage-driver-db="cadvisor" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204452 1977 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204455 1977 flags.go:64] FLAG: --storage-driver-password="root" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204458 1977 flags.go:64] FLAG: --storage-driver-secure="false" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204462 1977 flags.go:64] FLAG: --storage-driver-table="stats" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204465 1977 flags.go:64] FLAG: --storage-driver-user="root" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204468 1977 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204472 1977 flags.go:64] FLAG: --sync-frequency="1m0s" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204475 1977 flags.go:64] FLAG: --system-cgroups="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204478 1977 flags.go:64] FLAG: --system-reserved="cpu=500m,ephemeral-storage=1Gi,memory=1Gi" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204484 1977 flags.go:64] FLAG: --system-reserved-cgroup="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204488 1977 flags.go:64] FLAG: --tls-cert-file="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204492 1977 flags.go:64] FLAG: --tls-cipher-suites="[]" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204499 1977 flags.go:64] FLAG: --tls-min-version="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204502 1977 flags.go:64] FLAG: --tls-private-key-file="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204506 1977 flags.go:64] FLAG: --topology-manager-policy="none" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204509 1977 flags.go:64] FLAG: --topology-manager-scope="container" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204512 1977 flags.go:64] FLAG: --v="2" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204524 1977 flags.go:64] FLAG: --version="false" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204557 1977 flags.go:64] FLAG: --vmodule="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204566 1977 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204573 1977 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.204651 1977 feature_gate.go:246] feature gates: &{map[APIPriorityAndFairness:true CSIMigrationAzureFile:false CSIMigrationvSphere:false DownwardAPIHugePages:true RotateKubeletServerCertificate:true]} Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.209104 1977 server.go:413] "Kubelet version" kubeletVersion="v1.25.4+77bec7a" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.209122 1977 server.go:415] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.209192 1977 feature_gate.go:246] feature gates: &{map[APIPriorityAndFairness:true CSIMigrationAzureFile:false CSIMigrationvSphere:false DownwardAPIHugePages:true RotateKubeletServerCertificate:true]} Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.209290 1977 feature_gate.go:246] feature gates: &{map[APIPriorityAndFairness:true CSIMigrationAzureFile:false CSIMigrationvSphere:false DownwardAPIHugePages:true RotateKubeletServerCertificate:true]} Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.209451 1977 server.go:825] "Client rotation is on, will bootstrap in background" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.213776 1977 bootstrap.go:84] "Current kubeconfig file contents are still valid, no bootstrap necessary" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.213891 1977 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.214188 1977 server.go:882] "Starting client certificate rotation" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.214204 1977 certificate_manager.go:270] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.215679 1977 certificate_manager.go:270] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2023-01-06 06:46:04 +0000 UTC, rotation deadline is 2023-01-06 03:32:12.702966907 +0000 UTC Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.215702 1977 certificate_manager.go:270] kubernetes.io/kube-apiserver-client-kubelet: Waiting 9h45m36.4872673s for next certificate rotation Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.221605 1977 dynamic_cafile_content.go:119] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.221781 1977 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.223520 1977 manager.go:163] cAdvisor running in container: "/system.slice/kubelet.service" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.228603 1977 fs.go:133] Filesystem UUIDs: map[249B-6B8C:/dev/vda2 7e283ec8-4fd4-48d9-bedf-150e5fbbb374:/dev/vda3 fda260a9-581c-4391-b124-5cfcba02ef65:/dev/vda4] Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.228624 1977 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /sys/fs/cgroup:{mountpoint:/sys/fs/cgroup major:0 minor:25 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:43 fsType:tmpfs blockSize:0}] Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.228663 1977 nvidia.go:54] NVIDIA GPU metrics disabled Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.230606 1977 manager.go:212] Machine: {Timestamp:2023-01-05 17:46:36.230399395 +0000 UTC m=+0.317556024 CPUVendorID:GenuineIntel NumCores:2 NumPhysicalCores:1 NumSockets:1 CpuFrequency:2500000 MemoryCapacity:8192073728 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:2bf656612c3249b68a06a3657cb09103 SystemUUID:2bf65661-2c32-49b6-8a06-a3657cb09103 BootID:0795e9fa-b423-4c4c-88f3-736854a09041 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:4096036864 Type:vfs Inodes:1000009 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:4096036864 Type:vfs Inodes:1000009 HasInodes:true} {Device:/sys/fs/cgroup DeviceMajor:0 DeviceMinor:25 Capacity:4096036864 Type:vfs Inodes:1000009 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:128300593152 Type:vfs Inodes:62651840 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:43 Capacity:4096036864 Type:vfs Inodes:1000009 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:128849018880 Scheduler:mq-deadline}] NetworkDevices:[{Name:br0 MacAddress:6a:bb:63:d8:0d:48 Speed:0 Mtu:1450} {Name:ens5 MacAddress:00:16:3e:00:58:a6 Speed:-1 Mtu:1500} {Name:ovs-system MacAddress:72:d8:ef:62:c6:73 Speed:0 Mtu:1500} {Name:tun0 MacAddress:ee:58:26:93:e5:3c Speed:0 Mtu:1450} {Name:vxlan_sys_4789 MacAddress:56:c2:80:bc:2c:03 Speed:-1 Mtu:65000}] Topology:[{Id:0 Memory:8192073728 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0 1] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:1048576 Type:Unified Level:2}] UncoreCaches:[] SocketID:0}] Caches:[{Id:0 Size:37486592 Type:Unified Level:3}]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.231636 1977 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.231852 1977 manager.go:228] Version: {KernelVersion:4.18.0-372.39.1.el8_6.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 412.86.202212170457-0 (Ootpa) DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.234491 1977 container_manager_linux.go:262] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.234639 1977 container_manager_linux.go:267] "Creating Container Manager object based on Node Config" nodeConfig={RuntimeCgroupsName:/system.slice/crio.service SystemCgroupsName:/system.slice KubeletCgroupsName: KubeletOOMScoreAdj:-999 ContainerRuntime: CgroupsPerQOS:true CgroupRoot:/ CgroupDriver:systemd KubeletRootDir:/var/lib/kubelet ProtectKernelDefaults:false NodeAllocatableConfig:{KubeReservedCgroupName: SystemReservedCgroupName: ReservedSystemCPUs: EnforceNodeAllocatable:map[pods:{}] KubeReserved:map[] SystemReserved:map[cpu:{i:{value:500 scale:-3} d:{Dec:} s:500m Format:DecimalSI} ephemeral-storage:{i:{value:1073741824 scale:0} d:{Dec:} s:1Gi Format:BinarySI} memory:{i:{value:1073741824 scale:0} d:{Dec:} s:1Gi Format:BinarySI}] HardEvictionThresholds:[{Signal:nodefs.inodesFree Operator:LessThan Value:{Quantity: Percentage:0.05} GracePeriod:0s MinReclaim:} {Signal:imagefs.available Operator:LessThan Value:{Quantity: Percentage:0.15} GracePeriod:0s MinReclaim:} {Signal:memory.available Operator:LessThan Value:{Quantity:100Mi Percentage:0} GracePeriod:0s MinReclaim:} {Signal:nodefs.available Operator:LessThan Value:{Quantity: Percentage:0.1} GracePeriod:0s MinReclaim:}]} QOSReserved:map[] ExperimentalCPUManagerPolicy:none ExperimentalCPUManagerPolicyOptions:map[] ExperimentalTopologyManagerScope:container ExperimentalCPUManagerReconcilePeriod:10s ExperimentalMemoryManagerPolicy:None ExperimentalMemoryManagerReservedMemory:[] ExperimentalPodPidsLimit:4096 EnforceCPULimits:true CPUCFSQuotaPeriod:100ms ExperimentalTopologyManagerPolicy:none} Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.234673 1977 topology_manager.go:134] "Creating topology manager with policy per scope" topologyPolicyName="none" topologyScopeName="container" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.234694 1977 container_manager_linux.go:302] "Creating device plugin manager" devicePluginEnabled=true Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.235176 1977 manager.go:127] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.235201 1977 server.go:64] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.235361 1977 state_mem.go:36] "Initialized new in-memory state store" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.235436 1977 util_unix.go:104] "Using this endpoint is deprecated, please consider using full URL format" endpoint="/var/run/crio/crio.sock" URL="unix:///var/run/crio/crio.sock" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.243243 1977 remote_runtime.go:139] "Using CRI v1 runtime API" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.243268 1977 util_unix.go:104] "Using this endpoint is deprecated, please consider using full URL format" endpoint="/var/run/crio/crio.sock" URL="unix:///var/run/crio/crio.sock" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.264503 1977 remote_image.go:95] "Using CRI v1 image API" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.264524 1977 server.go:1136] "Using root directory" path="/var/lib/kubelet" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.266014 1977 kubelet.go:393] "Attempting to sync node with API server" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.266034 1977 kubelet.go:282] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.266072 1977 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.266087 1977 kubelet.go:293] "Adding apiserver pod source" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.266100 1977 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.268578 1977 kuberuntime_manager.go:240] "Container runtime initialized" containerRuntime="cri-o" version="1.25.1-5.rhaos4.12.git6005903.el8" apiVersion="v1" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.269437 1977 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.272473 1977 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/rbd" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.272498 1977 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/aws-ebs" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.272518 1977 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/gce-pd" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.272530 1977 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/cinder" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.272559 1977 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/azure-disk" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.272570 1977 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/azure-file" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.272581 1977 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/vsphere-volume" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.272602 1977 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.273052 1977 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.273072 1977 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.273084 1977 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.273102 1977 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.273114 1977 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/secret" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.273133 1977 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.273145 1977 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/glusterfs" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.273157 1977 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/cephfs" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.273168 1977 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.273184 1977 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/fc" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.273196 1977 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.273208 1977 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/projected" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.273219 1977 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.273869 1977 plugins.go:646] "Loaded volume plugin" pluginName="kubernetes.io/csi" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.274006 1977 server.go:1175] "Started kubelet" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl systemd[1]: Started Kubernetes Kubelet. Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: E0105 17:46:36.280299 1977 kubelet.go:1333] "Image garbage collection failed once. Stats initialization may not have completed yet" err="failed to get imageFs info: unable to find data in memory cache" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.282014 1977 certificate_manager.go:270] kubernetes.io/kubelet-serving: Certificate rotation is enabled Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.282045 1977 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.290632 1977 server.go:155] "Starting to listen" address="0.0.0.0" port=10250 Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.292090 1977 server.go:438] "Adding debug handlers to kubelet server" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.293028 1977 volume_manager.go:291] "The desired_state_of_world populator starts" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:36.296426310Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=914c6335-2455-40fa-8024-c06b41b81324 name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.299200 1977 volume_manager.go:293] "Starting Kubelet Volume Manager" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.300217 1977 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.293429 1977 certificate_manager.go:270] kubernetes.io/kubelet-serving: Certificate expiration is 2023-01-06 06:46:04 +0000 UTC, rotation deadline is 2023-01-06 04:22:32.706337533 +0000 UTC Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.300484 1977 certificate_manager.go:270] kubernetes.io/kubelet-serving: Waiting 10h35m56.40585947s for next certificate rotation Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:36.301532502Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=914c6335-2455-40fa-8024-c06b41b81324 name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.305652 1977 factory.go:153] Registering CRI-O factory Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.307269 1977 factory.go:55] Registering systemd factory Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.308069 1977 factory.go:103] Registering Raw factory Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.309769 1977 manager.go:1201] Started watching for new ooms in manager Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.311774 1977 manager.go:302] Starting recovery of all containers Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.412927 1977 kubelet_node_status.go:376] "Setting node annotation to enable volume controller attach/detach" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.459984 1977 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.460184 1977 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.460261 1977 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.460358 1977 kubelet_node_status.go:72] "Attempting to register node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.483973 1977 manager.go:307] Recovery completed Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.505139 1977 kubelet_node_status.go:110] "Node was previously registered" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.505429 1977 kubelet_node_status.go:75] "Successfully registered node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.524297 1977 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientMemory" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.524355 1977 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasNoDiskPressure" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.524370 1977 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeHasSufficientPID" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.524398 1977 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeNotReady" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.524578 1977 setters.go:545] "Node became not ready" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" condition={Type:Ready Status:False LastHeartbeatTime:2023-01-05 17:46:36.524377233 +0000 UTC m=+0.611533831 LastTransitionTime:2023-01-05 17:46:36.524377233 +0000 UTC m=+0.611533831 Reason:KubeletNotReady Message:[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]} Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.524608 1977 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeNotSchedulable" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.550562 1977 kubelet_network_linux.go:63] "Initialized iptables rules." protocol=IPv4 Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.570175 1977 cpu_manager.go:214] "Starting CPU manager" policy="none" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.570314 1977 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.570388 1977 state_mem.go:36] "Initialized new in-memory state store" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.571221 1977 policy_none.go:49] "None policy: Start" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.571977 1977 memory_manager.go:168] "Starting memorymanager" policy="None" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.572090 1977 state_mem.go:35] "Initializing new in-memory state store" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.573939 1977 container_manager_linux.go:427] "Updating kernel flag" flag="vm/overcommit_memory" expectedValue=1 actualValue=0 Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.574209 1977 container_manager_linux.go:427] "Updating kernel flag" flag="kernel/panic" expectedValue=10 actualValue=0 Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.614788 1977 manager.go:273] "Starting Device Plugin manager" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.620481 1977 manager.go:447] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.620605 1977 server.go:77] "Starting device plugin registration server" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.620872 1977 plugin_watcher.go:52] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.621021 1977 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.621088 1977 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: W0105 17:46:36.615033 1977 watcher.go:93] Error while processing event ("/sys/fs/cgroup/devices/kubepods.slice/kubepods-besteffort.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/devices/kubepods.slice/kubepods-besteffort.slice: no such file or directory Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.632357 1977 kubelet_network_linux.go:63] "Initialized iptables rules." protocol=IPv6 Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.632482 1977 status_manager.go:161] "Starting to sync pod status with apiserver" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.632582 1977 kubelet.go:2033] "Starting kubelet main sync loop" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: E0105 17:46:36.632678 1977 kubelet.go:2057] "Skipping pod synchronization" err="PLEG is not healthy: pleg has yet to be successful" Jan 05 17:46:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:36.733399 1977 kubelet.go:2119] "SyncLoop ADD" source="file" pods=[] Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.269266 1977 apiserver.go:52] "Watching apiserver" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.273572 1977 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-multus/network-metrics-daemon-hq594 openshift-multus/multus-4xhw7 openshift-ingress-canary/ingress-canary-8xrbh openshift-network-diagnostics/network-check-target-xmq2g openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4 openshift-dns/dns-default-2pfzf openshift-cluster-node-tuning-operator/tuned-smwj7 openshift-sdn/sdn-66nsp openshift-monitoring/node-exporter-2r6nf openshift-machine-config-operator/machine-config-daemon-hf8f5 openshift-multus/multus-additional-cni-plugins-5fgxc openshift-image-registry/node-ca-xw6d2 openshift-dns/node-resolver-nxzr8] Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.273645 1977 topology_manager.go:205] "Topology Admit Handler" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.273724 1977 topology_manager.go:205] "Topology Admit Handler" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.273781 1977 topology_manager.go:205] "Topology Admit Handler" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.273841 1977 topology_manager.go:205] "Topology Admit Handler" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.273964 1977 topology_manager.go:205] "Topology Admit Handler" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.274085 1977 topology_manager.go:205] "Topology Admit Handler" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.274214 1977 topology_manager.go:205] "Topology Admit Handler" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.274429 1977 topology_manager.go:205] "Topology Admit Handler" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.274588 1977 topology_manager.go:205] "Topology Admit Handler" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.274712 1977 topology_manager.go:205] "Topology Admit Handler" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.274862 1977 topology_manager.go:205] "Topology Admit Handler" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.274996 1977 topology_manager.go:205] "Topology Admit Handler" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.275123 1977 topology_manager.go:205] "Topology Admit Handler" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.307531 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-system-cni-dir\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.307606 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-sys\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.307636 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-run-netns\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.307666 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-run-ovs\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-ovs\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.307701 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-wtmp\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.307733 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-metrics-client-ca\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.307766 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfs2p\" (UniqueName: \"kubernetes.io/projected/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-kube-api-access-dfs2p\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.307795 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/659b8bb4-81b5-49bf-8127-1d326f30ba2f-serviceca\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.307826 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d5b0f5e6-3d70-4315-b9c5-1403036f517b-env-overrides\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.307858 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-run-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-kubernetes\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.307891 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-cnibin\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.307928 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-etc\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.307973 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmxtc\" (UniqueName: \"kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc\") pod \"network-check-target-xmq2g\" (UID: \"f364a949-09d7-4cfb-83ff-e532b822a557\") " pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.308015 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5b0f5e6-3d70-4315-b9c5-1403036f517b-config\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.308046 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-run-crio\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-crio\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.308076 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-lib-modules\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.308108 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wq4k7\" (UniqueName: \"kubernetes.io/projected/c0e47249-a5ff-460e-958d-53577387c154-kube-api-access-wq4k7\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.308144 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzfn5\" (UniqueName: \"kubernetes.io/projected/e04a47b4-aa8e-4d4e-9d45-aa8037dcb748-kube-api-access-lzfn5\") pod \"ingress-canary-8xrbh\" (UID: \"e04a47b4-aa8e-4d4e-9d45-aa8037dcb748\") " pod="openshift-ingress-canary/ingress-canary-8xrbh" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.308178 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-system-cni-dir\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.308210 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7xz4\" (UniqueName: \"kubernetes.io/projected/91265f01-9dcb-4046-961e-ad4c544224d7-kube-api-access-b7xz4\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.308243 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-device-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.308275 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g48vj\" (UniqueName: \"kubernetes.io/projected/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-kube-api-access-g48vj\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.308334 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-cnibin\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.308371 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xrks\" (UniqueName: \"kubernetes.io/projected/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-kube-api-access-5xrks\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.308403 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8f88cbe5-1221-4e6f-a6c9-f22da152b43f-hosts-file\") pod \"node-resolver-nxzr8\" (UID: \"8f88cbe5-1221-4e6f-a6c9-f22da152b43f\") " pod="openshift-dns/node-resolver-nxzr8" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.308438 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-host\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.308471 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-registration-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.308507 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"non-standard-root-system-trust-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bb84a70-232c-4467-ac0b-647030155908-non-standard-root-system-trust-ca-bundle\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.308558 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-mtu-migration\" (UniqueName: \"kubernetes.io/configmap/d5b0f5e6-3d70-4315-b9c5-1403036f517b-config-mtu-migration\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.308589 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-modules\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-modules\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.308620 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sdn-metrics-certs\" (UniqueName: \"kubernetes.io/secret/d5b0f5e6-3d70-4315-b9c5-1403036f517b-sdn-metrics-certs\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.308653 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx2gt\" (UniqueName: \"kubernetes.io/projected/50152296-9489-4fa8-aa42-7858debe1c08-kube-api-access-zx2gt\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.308697 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-sys\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.308732 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4sjn\" (UniqueName: \"kubernetes.io/projected/659b8bb4-81b5-49bf-8127-1d326f30ba2f-kube-api-access-h4sjn\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.308768 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-run-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-openshift-sdn\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.308800 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-plugin-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.308835 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-slash\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.308865 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-var-run-dbus\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.308895 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-os-release\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.308927 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-textfile\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.308958 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-rootfs\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.308991 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-proxy-tls\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.309024 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-tuned-profiles-data\" (UniqueName: \"kubernetes.io/configmap/91265f01-9dcb-4046-961e-ad4c544224d7-var-lib-tuned-profiles-data\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.309058 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-lib-cni-networks-openshift-sdn\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.309088 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-cni-binary-copy\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.309119 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-cni-bin\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.309150 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.309185 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c0e47249-a5ff-460e-958d-53577387c154-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.309215 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-etc-sysconfig\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.309245 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-root\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.309279 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zf2qr\" (UniqueName: \"kubernetes.io/projected/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-kube-api-access-zf2qr\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.309326 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.309364 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cookie-secret\" (UniqueName: \"kubernetes.io/secret/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-cookie-secret\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.309398 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-os-release\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.309431 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c0e47249-a5ff-460e-958d-53577387c154-cni-binary-copy\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.309467 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd-system\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-run-systemd-system\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.309498 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/659b8bb4-81b5-49bf-8127-1d326f30ba2f-host\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.309531 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-metrics-tls\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.309582 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzzxr\" (UniqueName: \"kubernetes.io/projected/1bb84a70-232c-4467-ac0b-647030155908-kube-api-access-vzzxr\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.309614 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-multus-cni-dir\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.309645 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-tls\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.309679 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.309711 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-kubelet-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.309742 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-dbus\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.309773 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nd4t\" (UniqueName: \"kubernetes.io/projected/d5b0f5e6-3d70-4315-b9c5-1403036f517b-kube-api-access-8nd4t\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.309806 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7www\" (UniqueName: \"kubernetes.io/projected/8f88cbe5-1221-4e6f-a6c9-f22da152b43f-kube-api-access-g7www\") pod \"node-resolver-nxzr8\" (UID: \"8f88cbe5-1221-4e6f-a6c9-f22da152b43f\") " pod="openshift-dns/node-resolver-nxzr8" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.309837 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-config-volume\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.309867 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-conf\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-cni-conf\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.309878 1977 reconciler.go:169] "Reconciler: start to sync state" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: W0105 17:46:37.382710 1977 watcher.go:93] Error while processing event ("/sys/fs/cgroup/devices/kubepods.slice/kubepods-burstable.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/devices/kubepods.slice/kubepods-burstable.slice: no such file or directory Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.418923 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-os-release\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.419046 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-textfile\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.419115 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-lib-cni-networks-openshift-sdn\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.419148 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-cni-binary-copy\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.419182 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-rootfs\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.419214 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-proxy-tls\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.419255 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"var-lib-tuned-profiles-data\" (UniqueName: \"kubernetes.io/configmap/91265f01-9dcb-4046-961e-ad4c544224d7-var-lib-tuned-profiles-data\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.419288 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-cni-bin\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.419344 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.419375 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-etc-sysconfig\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.419404 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-root\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.419443 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c0e47249-a5ff-460e-958d-53577387c154-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.419472 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/659b8bb4-81b5-49bf-8127-1d326f30ba2f-host\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.419504 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-metrics-tls\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.419534 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-zf2qr\" (UniqueName: \"kubernetes.io/projected/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-kube-api-access-zf2qr\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.419589 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.419621 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cookie-secret\" (UniqueName: \"kubernetes.io/secret/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-cookie-secret\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.419662 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-os-release\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.419695 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c0e47249-a5ff-460e-958d-53577387c154-cni-binary-copy\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.419731 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"run-systemd-system\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-run-systemd-system\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.419766 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-vzzxr\" (UniqueName: \"kubernetes.io/projected/1bb84a70-232c-4467-ac0b-647030155908-kube-api-access-vzzxr\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.419804 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-dbus\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.419836 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-8nd4t\" (UniqueName: \"kubernetes.io/projected/d5b0f5e6-3d70-4315-b9c5-1403036f517b-kube-api-access-8nd4t\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.419868 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-multus-cni-dir\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.419900 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-tls\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.419933 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.419965 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-kubelet-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.420002 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-config-volume\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.420033 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-cni-conf\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-cni-conf\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.420068 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-g7www\" (UniqueName: \"kubernetes.io/projected/8f88cbe5-1221-4e6f-a6c9-f22da152b43f-kube-api-access-g7www\") pod \"node-resolver-nxzr8\" (UID: \"8f88cbe5-1221-4e6f-a6c9-f22da152b43f\") " pod="openshift-dns/node-resolver-nxzr8" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.420099 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-system-cni-dir\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.420140 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-sys\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.420172 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/659b8bb4-81b5-49bf-8127-1d326f30ba2f-serviceca\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.420203 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d5b0f5e6-3d70-4315-b9c5-1403036f517b-env-overrides\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.420232 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-run-netns\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.420262 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-run-ovs\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-ovs\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.420295 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-wtmp\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.420339 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-metrics-client-ca\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.420373 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-dfs2p\" (UniqueName: \"kubernetes.io/projected/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-kube-api-access-dfs2p\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.420403 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5b0f5e6-3d70-4315-b9c5-1403036f517b-config\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.420434 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-run-crio\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-crio\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.420466 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-run-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-kubernetes\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.420498 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-cnibin\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.420528 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"etc\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-etc\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.420585 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-mmxtc\" (UniqueName: \"kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc\") pod \"network-check-target-xmq2g\" (UID: \"f364a949-09d7-4cfb-83ff-e532b822a557\") " pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.420641 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-wq4k7\" (UniqueName: \"kubernetes.io/projected/c0e47249-a5ff-460e-958d-53577387c154-kube-api-access-wq4k7\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.420675 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-lzfn5\" (UniqueName: \"kubernetes.io/projected/e04a47b4-aa8e-4d4e-9d45-aa8037dcb748-kube-api-access-lzfn5\") pod \"ingress-canary-8xrbh\" (UID: \"e04a47b4-aa8e-4d4e-9d45-aa8037dcb748\") " pod="openshift-ingress-canary/ingress-canary-8xrbh" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.420706 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-lib-modules\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.420741 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-g48vj\" (UniqueName: \"kubernetes.io/projected/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-kube-api-access-g48vj\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.420770 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-cnibin\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.420804 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-system-cni-dir\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.420838 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-b7xz4\" (UniqueName: \"kubernetes.io/projected/91265f01-9dcb-4046-961e-ad4c544224d7-kube-api-access-b7xz4\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.420873 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-device-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.420905 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config-mtu-migration\" (UniqueName: \"kubernetes.io/configmap/d5b0f5e6-3d70-4315-b9c5-1403036f517b-config-mtu-migration\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.420936 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-modules\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-modules\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.420971 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-5xrks\" (UniqueName: \"kubernetes.io/projected/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-kube-api-access-5xrks\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.421003 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8f88cbe5-1221-4e6f-a6c9-f22da152b43f-hosts-file\") pod \"node-resolver-nxzr8\" (UID: \"8f88cbe5-1221-4e6f-a6c9-f22da152b43f\") " pod="openshift-dns/node-resolver-nxzr8" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.421032 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-host\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.421068 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-registration-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.421105 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"non-standard-root-system-trust-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bb84a70-232c-4467-ac0b-647030155908-non-standard-root-system-trust-ca-bundle\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.421138 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-h4sjn\" (UniqueName: \"kubernetes.io/projected/659b8bb4-81b5-49bf-8127-1d326f30ba2f-kube-api-access-h4sjn\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.421171 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-var-run-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-openshift-sdn\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.421202 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"sdn-metrics-certs\" (UniqueName: \"kubernetes.io/secret/d5b0f5e6-3d70-4315-b9c5-1403036f517b-sdn-metrics-certs\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.421236 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-zx2gt\" (UniqueName: \"kubernetes.io/projected/50152296-9489-4fa8-aa42-7858debe1c08-kube-api-access-zx2gt\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.421267 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-sys\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.421298 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-slash\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.421352 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-var-run-dbus\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.421386 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"plugin-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-plugin-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.421688 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-os-release\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.421892 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-run-ovs\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-ovs\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.421929 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"plugin-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-plugin-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.422053 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-wtmp\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.422186 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-lib-cni-networks-openshift-sdn\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.422520 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-rootfs\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.422526 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-textfile\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.422977 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-metrics-client-ca\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.423109 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-run-crio\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-crio\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.423178 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-run-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-kubernetes\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.423247 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-cnibin\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.423349 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-cni-bin\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.423371 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5b0f5e6-3d70-4315-b9c5-1403036f517b-config\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.423995 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"var-lib-tuned-profiles-data\" (UniqueName: \"kubernetes.io/configmap/91265f01-9dcb-4046-961e-ad4c544224d7-var-lib-tuned-profiles-data\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.423995 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-lib-modules\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.424200 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-cnibin\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.424238 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-system-cni-dir\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.424468 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-device-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.424695 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config-mtu-migration\" (UniqueName: \"kubernetes.io/configmap/d5b0f5e6-3d70-4315-b9c5-1403036f517b-config-mtu-migration\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.424739 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-modules\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-modules\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.424975 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8f88cbe5-1221-4e6f-a6c9-f22da152b43f-hosts-file\") pod \"node-resolver-nxzr8\" (UID: \"8f88cbe5-1221-4e6f-a6c9-f22da152b43f\") " pod="openshift-dns/node-resolver-nxzr8" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.425023 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-host\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.425074 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-registration-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.425241 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"etc\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-etc\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.425348 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-etc-sysconfig\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.425387 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"root\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-root\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.425594 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/659b8bb4-81b5-49bf-8127-1d326f30ba2f-host\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.426065 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-run-openshift-sdn\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-openshift-sdn\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.429956 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-sys\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.430004 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-slash\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.430064 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-var-run-dbus\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.432115 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-os-release\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.432601 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"run-systemd-system\" (UniqueName: \"kubernetes.io/host-path/91265f01-9dcb-4046-961e-ad4c544224d7-run-systemd-system\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.432824 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-var-run-dbus\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-var-run-dbus\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.434065 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-multus-cni-dir\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.437335 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-proxy-tls\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.437671 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1bb84a70-232c-4467-ac0b-647030155908-kubelet-dir\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.437907 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c0e47249-a5ff-460e-958d-53577387c154-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.438068 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-cni-conf\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-cni-conf\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.438578 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-system-cni-dir\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.438953 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-sys\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.440777 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d5b0f5e6-3d70-4315-b9c5-1403036f517b-env-overrides\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.440821 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d5b0f5e6-3d70-4315-b9c5-1403036f517b-host-run-netns\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.442698 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c0e47249-a5ff-460e-958d-53577387c154-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.445464 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"non-standard-root-system-trust-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bb84a70-232c-4467-ac0b-647030155908-non-standard-root-system-trust-ca-bundle\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.446194 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-cni-binary-copy\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.446779 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"sdn-metrics-certs\" (UniqueName: \"kubernetes.io/secret/d5b0f5e6-3d70-4315-b9c5-1403036f517b-sdn-metrics-certs\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.447741 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c0e47249-a5ff-460e-958d-53577387c154-cni-binary-copy\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.448896 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/659b8bb4-81b5-49bf-8127-1d326f30ba2f-serviceca\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.449008 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-config-volume\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.449715 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.452769 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50152296-9489-4fa8-aa42-7858debe1c08-metrics-certs\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.452975 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cookie-secret\" (UniqueName: \"kubernetes.io/secret/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-cookie-secret\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.455667 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-node-exporter-tls\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.456833 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-metrics-tls\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.469756 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-wq4k7\" (UniqueName: \"kubernetes.io/projected/c0e47249-a5ff-460e-958d-53577387c154-kube-api-access-wq4k7\") pod \"multus-additional-cni-plugins-5fgxc\" (UID: \"c0e47249-a5ff-460e-958d-53577387c154\") " pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.470186 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4sjn\" (UniqueName: \"kubernetes.io/projected/659b8bb4-81b5-49bf-8127-1d326f30ba2f-kube-api-access-h4sjn\") pod \"node-ca-xw6d2\" (UID: \"659b8bb4-81b5-49bf-8127-1d326f30ba2f\") " pod="openshift-image-registry/node-ca-xw6d2" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.470697 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmxtc\" (UniqueName: \"kubernetes.io/projected/f364a949-09d7-4cfb-83ff-e532b822a557-kube-api-access-mmxtc\") pod \"network-check-target-xmq2g\" (UID: \"f364a949-09d7-4cfb-83ff-e532b822a557\") " pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.471031 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfs2p\" (UniqueName: \"kubernetes.io/projected/aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f-kube-api-access-dfs2p\") pod \"node-exporter-2r6nf\" (UID: \"aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f\") " pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.473704 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7xz4\" (UniqueName: \"kubernetes.io/projected/91265f01-9dcb-4046-961e-ad4c544224d7-kube-api-access-b7xz4\") pod \"tuned-smwj7\" (UID: \"91265f01-9dcb-4046-961e-ad4c544224d7\") " pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.476834 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzfn5\" (UniqueName: \"kubernetes.io/projected/e04a47b4-aa8e-4d4e-9d45-aa8037dcb748-kube-api-access-lzfn5\") pod \"ingress-canary-8xrbh\" (UID: \"e04a47b4-aa8e-4d4e-9d45-aa8037dcb748\") " pod="openshift-ingress-canary/ingress-canary-8xrbh" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.480655 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xrks\" (UniqueName: \"kubernetes.io/projected/997b7ab4-939e-465c-9c7d-4a2ebe3a797a-kube-api-access-5xrks\") pod \"machine-config-daemon-hf8f5\" (UID: \"997b7ab4-939e-465c-9c7d-4a2ebe3a797a\") " pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.482692 1977 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.483907243Z" level=info msg="Running pod sandbox: openshift-machine-config-operator/machine-config-daemon-hf8f5/POD" id=5dedd9e4-6079-44fd-b637-763716cf7c44 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.484242712Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.497354 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx2gt\" (UniqueName: \"kubernetes.io/projected/50152296-9489-4fa8-aa42-7858debe1c08-kube-api-access-zx2gt\") pod \"network-metrics-daemon-hq594\" (UID: \"50152296-9489-4fa8-aa42-7858debe1c08\") " pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.510628 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-g48vj\" (UniqueName: \"kubernetes.io/projected/1cc2aadb-14f2-4d6e-97fc-6fdab8889a63-kube-api-access-g48vj\") pod \"dns-default-2pfzf\" (UID: \"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63\") " pod="openshift-dns/dns-default-2pfzf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.510948 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-zf2qr\" (UniqueName: \"kubernetes.io/projected/84f9cd8a-5e58-4aff-aedd-906ed37dc97d-kube-api-access-zf2qr\") pod \"multus-4xhw7\" (UID: \"84f9cd8a-5e58-4aff-aedd-906ed37dc97d\") " pod="openshift-multus/multus-4xhw7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.511178 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nd4t\" (UniqueName: \"kubernetes.io/projected/d5b0f5e6-3d70-4315-b9c5-1403036f517b-kube-api-access-8nd4t\") pod \"sdn-66nsp\" (UID: \"d5b0f5e6-3d70-4315-b9c5-1403036f517b\") " pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.513411 1977 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.514334409Z" level=info msg="Running pod sandbox: openshift-multus/multus-additional-cni-plugins-5fgxc/POD" id=627bfb1c-0eb2-409a-ba2d-2fc13496d6e0 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.514423425Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.516472 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7www\" (UniqueName: \"kubernetes.io/projected/8f88cbe5-1221-4e6f-a6c9-f22da152b43f-kube-api-access-g7www\") pod \"node-resolver-nxzr8\" (UID: \"8f88cbe5-1221-4e6f-a6c9-f22da152b43f\") " pod="openshift-dns/node-resolver-nxzr8" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.516615 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzzxr\" (UniqueName: \"kubernetes.io/projected/1bb84a70-232c-4467-ac0b-647030155908-kube-api-access-vzzxr\") pod \"alibaba-disk-csi-driver-node-5sqb4\" (UID: \"1bb84a70-232c-4467-ac0b-647030155908\") " pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.527021144Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=5dedd9e4-6079-44fd-b637-763716cf7c44 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: W0105 17:46:37.534784 1977 watcher.go:93] Error while processing event ("/sys/fs/cgroup/devices/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod659b8bb4_81b5_49bf_8127_1d326f30ba2f.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/devices/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod659b8bb4_81b5_49bf_8127_1d326f30ba2f.slice: no such file or directory Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.535754 1977 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-xw6d2" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.540524796Z" level=info msg="Running pod sandbox: openshift-image-registry/node-ca-xw6d2/POD" id=327d3941-b490-4bad-8e5f-b9e4a841c606 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.541370708Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.553983299Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=627bfb1c-0eb2-409a-ba2d-2fc13496d6e0 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.564743886Z" level=info msg="Ran pod sandbox 403d4307d4ce0066eca971a22a71cc493d2fa1f3eced8bcfc2ebd58759cde28b with infra container: openshift-machine-config-operator/machine-config-daemon-hf8f5/POD" id=5dedd9e4-6079-44fd-b637-763716cf7c44 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: W0105 17:46:37.567326 1977 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod997b7ab4_939e_465c_9c7d_4a2ebe3a797a.slice/crio-403d4307d4ce0066eca971a22a71cc493d2fa1f3eced8bcfc2ebd58759cde28b.scope WatchSource:0}: Error finding container 403d4307d4ce0066eca971a22a71cc493d2fa1f3eced8bcfc2ebd58759cde28b: Status 404 returned error can't find the container with id 403d4307d4ce0066eca971a22a71cc493d2fa1f3eced8bcfc2ebd58759cde28b Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.569449 1977 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-nxzr8" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.572103900Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=327d3941-b490-4bad-8e5f-b9e4a841c606 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.574586340Z" level=info msg="Ran pod sandbox 0acf147f6c1cead086fc9a6bc2a1864cb87fecc6e2901c79505a3a3c2d4136d8 with infra container: openshift-multus/multus-additional-cni-plugins-5fgxc/POD" id=627bfb1c-0eb2-409a-ba2d-2fc13496d6e0 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.575459434Z" level=info msg="Running pod sandbox: openshift-dns/node-resolver-nxzr8/POD" id=40ffe60b-56c1-4aaf-b576-9c5741a205f2 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: W0105 17:46:37.576078 1977 watcher.go:93] Error while processing event ("/sys/fs/cgroup/devices/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0e47249_a5ff_460e_958d_53577387c154.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/devices/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0e47249_a5ff_460e_958d_53577387c154.slice: no such file or directory Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.576223358Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.590974089Z" level=info msg="Ran pod sandbox 66a75af31e155b3ae3bef9faad30f7bd130b43e5719f5074ae6205394f8db4f7 with infra container: openshift-image-registry/node-ca-xw6d2/POD" id=327d3941-b490-4bad-8e5f-b9e4a841c606 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.591160417Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:3fc008580e4e404960b2de4020a2939ab6421bdf8f51c0b60af4ad903272c741" id=1e4c1d82-b5d4-4c5c-9eaa-20d40dbcbaee name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.591472308Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cb9edf2251410c340a6125ff331924ab57a72ad3166bd9e1305b6d3030284990" id=169e0b57-673d-41dc-abdf-19880720e588 name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.595334225Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:6bb2a44a1f0506c18ef471cac7f7e8991d1e6b9d649ccbce50f751dd2863d06d,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:3fc008580e4e404960b2de4020a2939ab6421bdf8f51c0b60af4ad903272c741],Size_:382411122,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=1e4c1d82-b5d4-4c5c-9eaa-20d40dbcbaee name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.595658077Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8" id=a76fcd1d-b55d-4f32-aa82-f95693b94e9b name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: E0105 17:46:37.595711 1977 watcher.go:152] Failed to watch directory "/sys/fs/cgroup/devices/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod659b8bb4_81b5_49bf_8127_1d326f30ba2f.slice": open /sys/fs/cgroup/devices/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod659b8bb4_81b5_49bf_8127_1d326f30ba2f.slice: no such file or directory Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.598559138Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:c9e2954d8b9d9e1594d0a16b90086a9297ffd65773a6aa4a64f325e871ce83d1,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8],Size_:363009057,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=a76fcd1d-b55d-4f32-aa82-f95693b94e9b name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.605595706Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:3fc008580e4e404960b2de4020a2939ab6421bdf8f51c0b60af4ad903272c741" id=b798cff1-849c-4fe5-9085-569a8775c7a8 name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.605851177Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:6bb2a44a1f0506c18ef471cac7f7e8991d1e6b9d649ccbce50f751dd2863d06d,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:3fc008580e4e404960b2de4020a2939ab6421bdf8f51c0b60af4ad903272c741],Size_:382411122,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b798cff1-849c-4fe5-9085-569a8775c7a8 name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.605994875Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8" id=ee6b85ce-8e6b-4790-814e-aee34c301892 name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.606244099Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:c9e2954d8b9d9e1594d0a16b90086a9297ffd65773a6aa4a64f325e871ce83d1,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:37ffbf76753e8c339e5c1936f8f9a9e1d987574ab26f1b4228d69788e56525f8],Size_:363009057,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=ee6b85ce-8e6b-4790-814e-aee34c301892 name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.606244208Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2b208acf1f002c143fdb248dbee2517f7b6d168def4cfef1f3188939123d0dc3,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cb9edf2251410c340a6125ff331924ab57a72ad3166bd9e1305b6d3030284990],Size_:536350945,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=169e0b57-673d-41dc-abdf-19880720e588 name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.608209328Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=40ffe60b-56c1-4aaf-b576-9c5741a205f2 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.608669 1977 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-8xrbh" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.609683095Z" level=info msg="Creating container: openshift-image-registry/node-ca-xw6d2/node-ca" id=0d336939-5913-4e93-a8aa-72ddc80e1e82 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.610227167Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.609762481Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/egress-router-binary-copy" id=bd32f92b-c3d3-46e3-a46d-ec87e47354a7 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.613678759Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.609804077Z" level=info msg="Running pod sandbox: openshift-ingress-canary/ingress-canary-8xrbh/POD" id=a7bbab8d-0492-4b98-8549-e10134404e6a name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.613718847Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.609906463Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cb9edf2251410c340a6125ff331924ab57a72ad3166bd9e1305b6d3030284990" id=eb8ed7f3-c4f6-41e3-8afa-e1560cafaccb name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.621218 1977 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-2pfzf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.622464 1977 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/tuned-smwj7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.623826244Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2b208acf1f002c143fdb248dbee2517f7b6d168def4cfef1f3188939123d0dc3,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cb9edf2251410c340a6125ff331924ab57a72ad3166bd9e1305b6d3030284990],Size_:536350945,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=eb8ed7f3-c4f6-41e3-8afa-e1560cafaccb name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.624072093Z" level=info msg="Running pod sandbox: openshift-cluster-node-tuning-operator/tuned-smwj7/POD" id=4a38a01c-272b-40c7-9afb-eb2300cbcd75 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.624107804Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.624468353Z" level=info msg="Running pod sandbox: openshift-dns/dns-default-2pfzf/POD" id=7a11fa5d-3411-41a3-a3fe-df3e594289bf name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.624501889Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.632418367Z" level=info msg="Ran pod sandbox 09aeccea65d9703405b33f05b7cad2a5bb0bc55a0cbc0c82b5c23b0e4272ecaa with infra container: openshift-dns/node-resolver-nxzr8/POD" id=40ffe60b-56c1-4aaf-b576-9c5741a205f2 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.637735 1977 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-sdn/sdn-66nsp" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.640405255Z" level=info msg="Running pod sandbox: openshift-sdn/sdn-66nsp/POD" id=b849de3c-7c96-4800-b2ae-1d44d8e2dea7 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.640446129Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.640885488Z" level=info msg="Creating container: openshift-machine-config-operator/machine-config-daemon-hf8f5/machine-config-daemon" id=dfd60dec-70bb-4181-917f-7390be56a47c name=/runtime.v1.RuntimeService/CreateContainer Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.640975978Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.642103987Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f93a29f8d7455a5067ecde85394835bda877174193ab693b2b010d96c8cd9183" id=d0858f63-5d4f-44d8-958e-98fc27175ce1 name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.642377899Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:9b1e063a0faa2683ff00ee1f8b324af985c1c142530a63c7f8729ec8dab01b38,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f93a29f8d7455a5067ecde85394835bda877174193ab693b2b010d96c8cd9183],Size_:480731528,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=d0858f63-5d4f-44d8-958e-98fc27175ce1 name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.647730 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerStarted Data:0acf147f6c1cead086fc9a6bc2a1864cb87fecc6e2901c79505a3a3c2d4136d8} Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.647999 1977 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-4xhw7" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.648117567Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f93a29f8d7455a5067ecde85394835bda877174193ab693b2b010d96c8cd9183" id=433094ef-494a-42df-bed1-e9d3a8861ce4 name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.649458293Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:9b1e063a0faa2683ff00ee1f8b324af985c1c142530a63c7f8729ec8dab01b38,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f93a29f8d7455a5067ecde85394835bda877174193ab693b2b010d96c8cd9183],Size_:480731528,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=433094ef-494a-42df-bed1-e9d3a8861ce4 name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.650724812Z" level=info msg="Running pod sandbox: openshift-multus/multus-4xhw7/POD" id=0a6615ff-a9ed-42f9-924d-56fc5c01700b name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.650784537Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.652424 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" event=&{ID:997b7ab4-939e-465c-9c7d-4a2ebe3a797a Type:ContainerStarted Data:403d4307d4ce0066eca971a22a71cc493d2fa1f3eced8bcfc2ebd58759cde28b} Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.654698480Z" level=info msg="Creating container: openshift-dns/node-resolver-nxzr8/dns-node-resolver" id=3758c9ee-091d-4401-abb6-41c15bc25d44 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.654782279Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.657514 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-nxzr8" event=&{ID:8f88cbe5-1221-4e6f-a6c9-f22da152b43f Type:ContainerStarted Data:09aeccea65d9703405b33f05b7cad2a5bb0bc55a0cbc0c82b5c23b0e4272ecaa} Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.658560 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-xw6d2" event=&{ID:659b8bb4-81b5-49bf-8127-1d326f30ba2f Type:ContainerStarted Data:66a75af31e155b3ae3bef9faad30f7bd130b43e5719f5074ae6205394f8db4f7} Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.666380 1977 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.666708756Z" level=info msg="Running pod sandbox: openshift-network-diagnostics/network-check-target-xmq2g/POD" id=84839fef-8cc1-44ae-9dfd-90086ea73636 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.666764529Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.679754729Z" level=info msg="Got pod network &{Name:ingress-canary-8xrbh Namespace:openshift-ingress-canary ID:8510bb13dab2503578d194877e82efbb3c57ad0a439fe2f087985adeeca2f21b UID:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748 NetNS:/var/run/netns/3f4b8877-65b1-4936-8a9d-2522145b5205 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.679779493Z" level=info msg="Adding pod openshift-ingress-canary_ingress-canary-8xrbh to CNI network \"multus-cni-network\" (type=multus)" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.686913 1977 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.687263758Z" level=info msg="Running pod sandbox: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/POD" id=28f7f763-2201-4626-9efc-23ed6f5524d6 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.687315494Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.689677212Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=4a38a01c-272b-40c7-9afb-eb2300cbcd75 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.702960 1977 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hq594" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.703910675Z" level=info msg="Running pod sandbox: openshift-multus/network-metrics-daemon-hq594/POD" id=3b85e0f9-6bf9-45bc-b186-ac78e116f33e name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.703951979Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: W0105 17:46:37.720037 1977 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91265f01_9dcb_4046_961e_ad4c544224d7.slice/crio-2c86ae5de21a07ed9744965ec30c6937bb991497b03314f8aa2a37768c7972e0.scope WatchSource:0}: Error finding container 2c86ae5de21a07ed9744965ec30c6937bb991497b03314f8aa2a37768c7972e0: Status 404 returned error can't find the container with id 2c86ae5de21a07ed9744965ec30c6937bb991497b03314f8aa2a37768c7972e0 Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.726261476Z" level=info msg="Got pod network &{Name:network-check-target-xmq2g Namespace:openshift-network-diagnostics ID:214a9c55e49c804acc78be4ce16b5a653ad517d1495dac27266eba0b4f45de68 UID:f364a949-09d7-4cfb-83ff-e532b822a557 NetNS:/var/run/netns/53f617ed-730b-4fb4-abb1-c8d0d0bdef7e Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.726285820Z" level=info msg="Adding pod openshift-network-diagnostics_network-check-target-xmq2g to CNI network \"multus-cni-network\" (type=multus)" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.729760999Z" level=info msg="Ran pod sandbox 2c86ae5de21a07ed9744965ec30c6937bb991497b03314f8aa2a37768c7972e0 with infra container: openshift-cluster-node-tuning-operator/tuned-smwj7/POD" id=4a38a01c-272b-40c7-9afb-eb2300cbcd75 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:37.731707 1977 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-2r6nf" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.732660825Z" level=info msg="Running pod sandbox: openshift-monitoring/node-exporter-2r6nf/POD" id=6b38ccfb-da8f-4523-a7f3-d9e15c837885 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.732697867Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.732990328Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ca6ba83315a4882f5c6c4e01c79238e1b794b0e8bfdc2f004990f516e8e0b056" id=21a4b503-e590-4f91-8cd9-eeee8c9733fb name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.737296397Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b2f8382287975a85853941eec2c606723c2ed1c24f59fa46ebdb470b91a481f5,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ca6ba83315a4882f5c6c4e01c79238e1b794b0e8bfdc2f004990f516e8e0b056],Size_:505526574,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=21a4b503-e590-4f91-8cd9-eeee8c9733fb name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.765960820Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ca6ba83315a4882f5c6c4e01c79238e1b794b0e8bfdc2f004990f516e8e0b056" id=08cbf2b0-a92d-4744-966e-75101357624a name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.766175554Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b2f8382287975a85853941eec2c606723c2ed1c24f59fa46ebdb470b91a481f5,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ca6ba83315a4882f5c6c4e01c79238e1b794b0e8bfdc2f004990f516e8e0b056],Size_:505526574,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=08cbf2b0-a92d-4744-966e-75101357624a name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.767398264Z" level=info msg="Creating container: openshift-cluster-node-tuning-operator/tuned-smwj7/tuned" id=9b6daba5-5d2f-491a-b2f3-14a0b0d533fa name=/runtime.v1.RuntimeService/CreateContainer Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.767507340Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.778026004Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=b849de3c-7c96-4800-b2ae-1d44d8e2dea7 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.825279872Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=0a6615ff-a9ed-42f9-924d-56fc5c01700b name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: W0105 17:46:37.831751 1977 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5b0f5e6_3d70_4315_b9c5_1403036f517b.slice/crio-00dc400e22262c52c28041ed9089f2331218b00afa3df045b7cd42b1d70bf3fb.scope WatchSource:0}: Error finding container 00dc400e22262c52c28041ed9089f2331218b00afa3df045b7cd42b1d70bf3fb: Status 404 returned error can't find the container with id 00dc400e22262c52c28041ed9089f2331218b00afa3df045b7cd42b1d70bf3fb Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.838515914Z" level=info msg="Ran pod sandbox 00dc400e22262c52c28041ed9089f2331218b00afa3df045b7cd42b1d70bf3fb with infra container: openshift-sdn/sdn-66nsp/POD" id=b849de3c-7c96-4800-b2ae-1d44d8e2dea7 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: W0105 17:46:37.843081 1977 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod84f9cd8a_5e58_4aff_aedd_906ed37dc97d.slice/crio-c2921d3b0822033449151099f7e4a989b1b0cb7526ca50c89a34c94fc6dd7157.scope WatchSource:0}: Error finding container c2921d3b0822033449151099f7e4a989b1b0cb7526ca50c89a34c94fc6dd7157: Status 404 returned error can't find the container with id c2921d3b0822033449151099f7e4a989b1b0cb7526ca50c89a34c94fc6dd7157 Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.843838340Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:30387b6f2e605bb8d018e0c09f9d0275048fc5e7966137fbe6bd4307c88d3fdd" id=59baff92-1d44-411f-904b-a2e54fe662a1 name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.844122825Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:cff39d57d2c205a584f8bf4299b1e125fa1a96ac92f680117ba82c20334754d1,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:30387b6f2e605bb8d018e0c09f9d0275048fc5e7966137fbe6bd4307c88d3fdd],Size_:701241664,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=59baff92-1d44-411f-904b-a2e54fe662a1 name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.849848179Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:30387b6f2e605bb8d018e0c09f9d0275048fc5e7966137fbe6bd4307c88d3fdd" id=0dd38af1-70b5-43ad-bf1d-7a24152c381a name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.850044200Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:cff39d57d2c205a584f8bf4299b1e125fa1a96ac92f680117ba82c20334754d1,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:30387b6f2e605bb8d018e0c09f9d0275048fc5e7966137fbe6bd4307c88d3fdd],Size_:701241664,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=0dd38af1-70b5-43ad-bf1d-7a24152c381a name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.855987307Z" level=info msg="Ran pod sandbox c2921d3b0822033449151099f7e4a989b1b0cb7526ca50c89a34c94fc6dd7157 with infra container: openshift-multus/multus-4xhw7/POD" id=0a6615ff-a9ed-42f9-924d-56fc5c01700b name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.860961302Z" level=info msg="Creating container: openshift-sdn/sdn-66nsp/sdn" id=9fb72343-a92e-4a87-a72c-2cbd4360358e name=/runtime.v1.RuntimeService/CreateContainer Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.862408453Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.865622517Z" level=info msg="Got pod network &{Name:network-metrics-daemon-hq594 Namespace:openshift-multus ID:7e8ab5f06c8a76365c09171cecfc24f31a6a7f245ad42d5208ec29334d98b2c0 UID:50152296-9489-4fa8-aa42-7858debe1c08 NetNS:/var/run/netns/ba9d8790-155d-4a0b-bb29-8fbef77d7fe5 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.865731418Z" level=info msg="Adding pod openshift-multus_network-metrics-daemon-hq594 to CNI network \"multus-cni-network\" (type=multus)" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.867636568Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=b803d419-fa0a-4184-a371-93e6cbe32332 name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.874794164Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=28f7f763-2201-4626-9efc-23ed6f5524d6 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.876229566Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b803d419-fa0a-4184-a371-93e6cbe32332 name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.890959515Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=c6f77a56-df88-4e75-bbb8-d3d00c01dfa3 name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.909166103Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=c6f77a56-df88-4e75-bbb8-d3d00c01dfa3 name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: W0105 17:46:37.925654 1977 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1bb84a70_232c_4467_ac0b_647030155908.slice/crio-d7e9f4289f7cd500918dbb78134ef857afc6259662f81e9e13bb19cba691b342.scope WatchSource:0}: Error finding container d7e9f4289f7cd500918dbb78134ef857afc6259662f81e9e13bb19cba691b342: Status 404 returned error can't find the container with id d7e9f4289f7cd500918dbb78134ef857afc6259662f81e9e13bb19cba691b342 Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.928016472Z" level=info msg="Creating container: openshift-multus/multus-4xhw7/kube-multus" id=6bebba01-6863-4df2-81ab-da3fa0074273 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.928115127Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.937800024Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=6b38ccfb-da8f-4523-a7f3-d9e15c837885 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.943275140Z" level=info msg="Ran pod sandbox d7e9f4289f7cd500918dbb78134ef857afc6259662f81e9e13bb19cba691b342 with infra container: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/POD" id=28f7f763-2201-4626-9efc-23ed6f5524d6 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.952804874Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7" id=6cc4cf5c-240d-4bdc-aee8-2498a2006d9c name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:37.975270245Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:6e78a44a973f2298da6f51c1418094403ee8c6c2b57bde9d3285b9e2ac2ba129,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7],Size_:397889589,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=6cc4cf5c-240d-4bdc-aee8-2498a2006d9c name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: W0105 17:46:37.991401 1977 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa84cbad_3d0f_43bd_a8a9_46dc1cd9c57f.slice/crio-bccfa67f1ae302d09897f752660b0b20a0bf1f6fa4b4d3a20946611940b2e44e.scope WatchSource:0}: Error finding container bccfa67f1ae302d09897f752660b0b20a0bf1f6fa4b4d3a20946611940b2e44e: Status 404 returned error can't find the container with id bccfa67f1ae302d09897f752660b0b20a0bf1f6fa4b4d3a20946611940b2e44e Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.015145128Z" level=info msg="Ran pod sandbox bccfa67f1ae302d09897f752660b0b20a0bf1f6fa4b4d3a20946611940b2e44e with infra container: openshift-monitoring/node-exporter-2r6nf/POD" id=6b38ccfb-da8f-4523-a7f3-d9e15c837885 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.017770227Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7" id=8db7388b-6407-460a-9428-aa6353ac7f30 name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.018025753Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:6e78a44a973f2298da6f51c1418094403ee8c6c2b57bde9d3285b9e2ac2ba129,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:542a2ddcb0c87d58055f8c7044d2af8278115088b88dfa45c2fa9399570d14a7],Size_:397889589,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=8db7388b-6407-460a-9428-aa6353ac7f30 name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.052809109Z" level=info msg="Creating container: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-driver" id=0ccb5e8e-3660-4d11-924b-aa20171551ac name=/runtime.v1.RuntimeService/CreateContainer Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.052925256Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.067602318Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180" id=3d36d0f2-b003-44e2-bff4-9ba21ed5bcc7 name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.107848731Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b4cde205857c1e0a984b38c61275a8370136432c7c84dbfbd7f4835f26bbb935,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180],Size_:297268865,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=3d36d0f2-b003-44e2-bff4-9ba21ed5bcc7 name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.126825932Z" level=info msg="Created container 0c02c3dac58dc1adc2307d741bdaf1b4e1bdeb4e549f9ca7ec6e7d7d8d92dd23: openshift-sdn/sdn-66nsp/sdn" id=9fb72343-a92e-4a87-a72c-2cbd4360358e name=/runtime.v1.RuntimeService/CreateContainer Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.127773602Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180" id=a2ed876c-bf74-4500-a706-6e01662b9b02 name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.128097301Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b4cde205857c1e0a984b38c61275a8370136432c7c84dbfbd7f4835f26bbb935,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180],Size_:297268865,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=a2ed876c-bf74-4500-a706-6e01662b9b02 name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.134787209Z" level=info msg="Creating container: openshift-monitoring/node-exporter-2r6nf/init-textfile" id=f21b3cc5-5948-40bf-8485-a6acc0b6bd17 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.134904169Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.135433789Z" level=info msg="Created container 5487557a3b96f175c86652592000b00e0cceb48a7bd3adcdaf1bab23321f7bf7: openshift-cluster-node-tuning-operator/tuned-smwj7/tuned" id=9b6daba5-5d2f-491a-b2f3-14a0b0d533fa name=/runtime.v1.RuntimeService/CreateContainer Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.136333218Z" level=info msg="Created container bc2cc512d092efd272d111021170640072857278b10f2d60e14eb2e45afecd38: openshift-machine-config-operator/machine-config-daemon-hf8f5/machine-config-daemon" id=dfd60dec-70bb-4181-917f-7390be56a47c name=/runtime.v1.RuntimeService/CreateContainer Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.138432576Z" level=info msg="Starting container: 0c02c3dac58dc1adc2307d741bdaf1b4e1bdeb4e549f9ca7ec6e7d7d8d92dd23" id=a62658cb-7255-445c-8dd2-e3cdb1f4e4e1 name=/runtime.v1.RuntimeService/StartContainer Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.151434738Z" level=info msg="Starting container: bc2cc512d092efd272d111021170640072857278b10f2d60e14eb2e45afecd38" id=0af4961d-616d-4dc3-af07-f376511de7e8 name=/runtime.v1.RuntimeService/StartContainer Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.161362388Z" level=info msg="Starting container: 5487557a3b96f175c86652592000b00e0cceb48a7bd3adcdaf1bab23321f7bf7" id=9a284863-66e0-4340-871b-b6688934a715 name=/runtime.v1.RuntimeService/StartContainer Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.185365972Z" level=info msg="Started container" PID=2186 containerID=0c02c3dac58dc1adc2307d741bdaf1b4e1bdeb4e549f9ca7ec6e7d7d8d92dd23 description=openshift-sdn/sdn-66nsp/sdn id=a62658cb-7255-445c-8dd2-e3cdb1f4e4e1 name=/runtime.v1.RuntimeService/StartContainer sandboxID=00dc400e22262c52c28041ed9089f2331218b00afa3df045b7cd42b1d70bf3fb Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.221597736Z" level=info msg="Started container" PID=2185 containerID=5487557a3b96f175c86652592000b00e0cceb48a7bd3adcdaf1bab23321f7bf7 description=openshift-cluster-node-tuning-operator/tuned-smwj7/tuned id=9a284863-66e0-4340-871b-b6688934a715 name=/runtime.v1.RuntimeService/StartContainer sandboxID=2c86ae5de21a07ed9744965ec30c6937bb991497b03314f8aa2a37768c7972e0 Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.240429656Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=03f02a87-2ef3-4d81-9361-801f83449858 name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.240669537Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=03f02a87-2ef3-4d81-9361-801f83449858 name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.247843091Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=1fa2a290-01d8-4244-85ee-6e340de1b72d name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.248189311Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=1fa2a290-01d8-4244-85ee-6e340de1b72d name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.248604112Z" level=info msg="Started container" PID=2197 containerID=bc2cc512d092efd272d111021170640072857278b10f2d60e14eb2e45afecd38 description=openshift-machine-config-operator/machine-config-daemon-hf8f5/machine-config-daemon id=0af4961d-616d-4dc3-af07-f376511de7e8 name=/runtime.v1.RuntimeService/StartContainer sandboxID=403d4307d4ce0066eca971a22a71cc493d2fa1f3eced8bcfc2ebd58759cde28b Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.268754449Z" level=info msg="Creating container: openshift-sdn/sdn-66nsp/kube-rbac-proxy" id=cb80297b-ba12-408c-9d0b-3da5c19bd24d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.268869482Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.275266037Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.360633423Z" level=info msg="Created container 98021e0bcc407387017be2cd1113ad3d2304878175ad6091964550330c9401a5: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-driver" id=0ccb5e8e-3660-4d11-924b-aa20171551ac name=/runtime.v1.RuntimeService/CreateContainer Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.362006444Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.362155796Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.373293170Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.374001747Z" level=info msg="Starting container: 98021e0bcc407387017be2cd1113ad3d2304878175ad6091964550330c9401a5" id=64e2d89d-ddbe-4d2b-a254-998e371d886d name=/runtime.v1.RuntimeService/StartContainer Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.453638065Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=f1755244-7468-4c57-8bf1-62f7907a38b3 name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.454090546Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=f1755244-7468-4c57-8bf1-62f7907a38b3 name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.456902597Z" level=info msg="Started container" PID=2255 containerID=98021e0bcc407387017be2cd1113ad3d2304878175ad6091964550330c9401a5 description=openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-driver id=64e2d89d-ddbe-4d2b-a254-998e371d886d name=/runtime.v1.RuntimeService/StartContainer sandboxID=d7e9f4289f7cd500918dbb78134ef857afc6259662f81e9e13bb19cba691b342 Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.479851278Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.482074612Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.484234126Z" level=info msg="CNI monitoring event WRITE \"/var/lib/cni/bin/openshift-sdn\"" Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.485371822Z" level=info msg="Created container 241a7e1ea7bdaf847563b2fe6abacb680c8276e1479eff398b9a37196ef11530: openshift-monitoring/node-exporter-2r6nf/init-textfile" id=f21b3cc5-5948-40bf-8485-a6acc0b6bd17 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.485688099Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=fdca1dd1-59a5-4a49-bf6f-6fa0f4c41c8a name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.486253925Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=fdca1dd1-59a5-4a49-bf6f-6fa0f4c41c8a name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.513069018Z" level=info msg="Creating container: openshift-machine-config-operator/machine-config-daemon-hf8f5/oauth-proxy" id=862b9ab0-df5b-428b-84da-56241504c3c0 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.513211056Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.518811309Z" level=info msg="Starting container: 241a7e1ea7bdaf847563b2fe6abacb680c8276e1479eff398b9a37196ef11530" id=fa0dc8bb-5b73-45b0-8556-f74d4945b0ef name=/runtime.v1.RuntimeService/StartContainer Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.614501253Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:72c764b55efc147864960930ae05b7928c5f976eb76820b38997f0df7afab820" id=b2fa1637-35d5-49bf-9b43-b80aa5aeb13b name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.614752244Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:0bb939dc4e915f972f52c355cea411db098a9c605b2ca4bc1048447ae297391e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:72c764b55efc147864960930ae05b7928c5f976eb76820b38997f0df7afab820],Size_:293414029,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b2fa1637-35d5-49bf-9b43-b80aa5aeb13b name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.623733295Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:72c764b55efc147864960930ae05b7928c5f976eb76820b38997f0df7afab820" id=cfde60c8-8b7b-4b59-a95b-559cc0e8482f name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.624848490Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:0bb939dc4e915f972f52c355cea411db098a9c605b2ca4bc1048447ae297391e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:72c764b55efc147864960930ae05b7928c5f976eb76820b38997f0df7afab820],Size_:293414029,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=cfde60c8-8b7b-4b59-a95b-559cc0e8482f name=/runtime.v1.ImageService/ImageStatus Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.649491250Z" level=info msg="Creating container: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-node-driver-registrar" id=c470ecd3-3cdd-4f07-b684-086994cae28d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.649631196Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:38.681800 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" event=&{ID:997b7ab4-939e-465c-9c7d-4a2ebe3a797a Type:ContainerStarted Data:bc2cc512d092efd272d111021170640072857278b10f2d60e14eb2e45afecd38} Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:38.686396 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4xhw7" event=&{ID:84f9cd8a-5e58-4aff-aedd-906ed37dc97d Type:ContainerStarted Data:c2921d3b0822033449151099f7e4a989b1b0cb7526ca50c89a34c94fc6dd7157} Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:38.687612 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-smwj7" event=&{ID:91265f01-9dcb-4046-961e-ad4c544224d7 Type:ContainerStarted Data:5487557a3b96f175c86652592000b00e0cceb48a7bd3adcdaf1bab23321f7bf7} Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:38.687641 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-smwj7" event=&{ID:91265f01-9dcb-4046-961e-ad4c544224d7 Type:ContainerStarted Data:2c86ae5de21a07ed9744965ec30c6937bb991497b03314f8aa2a37768c7972e0} Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:38.689756 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-2r6nf" event=&{ID:aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f Type:ContainerStarted Data:bccfa67f1ae302d09897f752660b0b20a0bf1f6fa4b4d3a20946611940b2e44e} Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:38.696016 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" event=&{ID:1bb84a70-232c-4467-ac0b-647030155908 Type:ContainerStarted Data:98021e0bcc407387017be2cd1113ad3d2304878175ad6091964550330c9401a5} Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:38.696049 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" event=&{ID:1bb84a70-232c-4467-ac0b-647030155908 Type:ContainerStarted Data:d7e9f4289f7cd500918dbb78134ef857afc6259662f81e9e13bb19cba691b342} Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:38.711021 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-sdn/sdn-66nsp" event=&{ID:d5b0f5e6-3d70-4315-b9c5-1403036f517b Type:ContainerStarted Data:0c02c3dac58dc1adc2307d741bdaf1b4e1bdeb4e549f9ca7ec6e7d7d8d92dd23} Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 17:46:38.711050 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-sdn/sdn-66nsp" event=&{ID:d5b0f5e6-3d70-4315-b9c5-1403036f517b Type:ContainerStarted Data:00dc400e22262c52c28041ed9089f2331218b00afa3df045b7cd42b1d70bf3fb} Jan 05 17:46:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:38.961754097Z" level=info msg="Started container" PID=2309 containerID=241a7e1ea7bdaf847563b2fe6abacb680c8276e1479eff398b9a37196ef11530 description=openshift-monitoring/node-exporter-2r6nf/init-textfile id=fa0dc8bb-5b73-45b0-8556-f74d4945b0ef name=/runtime.v1.RuntimeService/StartContainer sandboxID=bccfa67f1ae302d09897f752660b0b20a0bf1f6fa4b4d3a20946611940b2e44e Jan 05 17:46:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:39.046380704Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 17:46:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 17:46:39.047482302Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:46:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:40.392466191Z" level=info msg="Created container 4ea62ec0ac85653a3c58bb51ccde4d6829ab49156353196af96d1e63801074da: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-node-driver-registrar" id=c470ecd3-3cdd-4f07-b684-086994cae28d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:40.442677 1977 generic.go:296] "Generic (PLEG): container finished" podID=aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f containerID="241a7e1ea7bdaf847563b2fe6abacb680c8276e1479eff398b9a37196ef11530" exitCode=0 Jan 05 09:46:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:40.444089932Z" level=info msg="Starting container: 4ea62ec0ac85653a3c58bb51ccde4d6829ab49156353196af96d1e63801074da" id=5abb5e7f-6300-40e0-8acf-9a4415d9c81a name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:46:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:40.461480131Z" level=info msg="Created container ce8da87ec13d04b847764bc661e8abc9371a4ebe65e568a5847aadad12983ad3: openshift-sdn/sdn-66nsp/kube-rbac-proxy" id=cb80297b-ba12-408c-9d0b-3da5c19bd24d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:40.445640234Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180" id=b6bddc3c-79bc-4e4d-bfb0-ba8052ac20c5 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:40.463441227Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b4cde205857c1e0a984b38c61275a8370136432c7c84dbfbd7f4835f26bbb935,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180],Size_:297268865,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b6bddc3c-79bc-4e4d-bfb0-ba8052ac20c5 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:40.444401 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-2r6nf" event=&{ID:aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f Type:ContainerDied Data:241a7e1ea7bdaf847563b2fe6abacb680c8276e1479eff398b9a37196ef11530} Jan 05 09:46:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:40.511195347Z" level=info msg="Starting container: ce8da87ec13d04b847764bc661e8abc9371a4ebe65e568a5847aadad12983ad3" id=ee44be43-a372-42ae-8b26-681959c3f10e name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:46:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:40.537088902Z" level=info msg="Started container" PID=2672 containerID=4ea62ec0ac85653a3c58bb51ccde4d6829ab49156353196af96d1e63801074da description=openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-node-driver-registrar id=5abb5e7f-6300-40e0-8acf-9a4415d9c81a name=/runtime.v1.RuntimeService/StartContainer sandboxID=d7e9f4289f7cd500918dbb78134ef857afc6259662f81e9e13bb19cba691b342 Jan 05 09:46:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:40.538731435Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180" id=c68a0b36-c240-470c-b170-4e2e5c9382dd name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:40.539391362Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b4cde205857c1e0a984b38c61275a8370136432c7c84dbfbd7f4835f26bbb935,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:c4218978f242fbc800301ede5f4a53c0520dbcccc0b65d88c653731aa2781180],Size_:297268865,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=c68a0b36-c240-470c-b170-4e2e5c9382dd name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:40.553719556Z" level=info msg="Creating container: openshift-monitoring/node-exporter-2r6nf/node-exporter" id=a78e8163-64b1-4755-b378-35ae8d60bba0 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:40.553823738Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:46:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:40.614921268Z" level=info msg="Created container db2c1a2ca7fd456ff77bdd0c3b85bc2af46b67cf77238b4d8471a3e5d72799d6: openshift-machine-config-operator/machine-config-daemon-hf8f5/oauth-proxy" id=862b9ab0-df5b-428b-84da-56241504c3c0 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:40.617496583Z" level=info msg="Starting container: db2c1a2ca7fd456ff77bdd0c3b85bc2af46b67cf77238b4d8471a3e5d72799d6" id=0e221af7-773f-4822-81c2-796deb4cc082 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:46:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:40.647034935Z" level=info msg="Started container" PID=2615 containerID=ce8da87ec13d04b847764bc661e8abc9371a4ebe65e568a5847aadad12983ad3 description=openshift-sdn/sdn-66nsp/kube-rbac-proxy id=ee44be43-a372-42ae-8b26-681959c3f10e name=/runtime.v1.RuntimeService/StartContainer sandboxID=00dc400e22262c52c28041ed9089f2331218b00afa3df045b7cd42b1d70bf3fb Jan 05 09:46:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:40.681024802Z" level=info msg="Started container" PID=2682 containerID=db2c1a2ca7fd456ff77bdd0c3b85bc2af46b67cf77238b4d8471a3e5d72799d6 description=openshift-machine-config-operator/machine-config-daemon-hf8f5/oauth-proxy id=0e221af7-773f-4822-81c2-796deb4cc082 name=/runtime.v1.RuntimeService/StartContainer sandboxID=403d4307d4ce0066eca971a22a71cc493d2fa1f3eced8bcfc2ebd58759cde28b Jan 05 09:46:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:40.689832862Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ea25c2f3749dee6981359b4655a9acfcd4d95547066ccdf328354b548da74932" id=7e52682b-aa03-4b2a-8dbf-ef3e2fb20395 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:40.690418447Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:58a5591f9334b70902e27f9a172f3e141b8f07a879507b850a76c2499014eec9,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ea25c2f3749dee6981359b4655a9acfcd4d95547066ccdf328354b548da74932],Size_:291396124,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=7e52682b-aa03-4b2a-8dbf-ef3e2fb20395 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:40.699201920Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ea25c2f3749dee6981359b4655a9acfcd4d95547066ccdf328354b548da74932" id=be6cf1aa-32ba-40cb-b886-92b7f0d1523d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:40.699413649Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:58a5591f9334b70902e27f9a172f3e141b8f07a879507b850a76c2499014eec9,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ea25c2f3749dee6981359b4655a9acfcd4d95547066ccdf328354b548da74932],Size_:291396124,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=be6cf1aa-32ba-40cb-b886-92b7f0d1523d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:40.735026738Z" level=info msg="Creating container: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-liveness-probe" id=978b41d1-3413-4f50-a5d1-ada60633b9c4 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:40.735174783Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:46:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:40.812247505Z" level=info msg="Created container d628617bbda48d44a43a225bf2f5e855121106f3afd330a3fdcb502bb0180e53: openshift-multus/multus-4xhw7/kube-multus" id=6bebba01-6863-4df2-81ab-da3fa0074273 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:40.825958899Z" level=info msg="Starting container: d628617bbda48d44a43a225bf2f5e855121106f3afd330a3fdcb502bb0180e53" id=91a9e4f8-bcd8-4039-9807-463f572690b5 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:46:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:40.904260030Z" level=info msg="Created container ffe41dec05c7f28d2bbc3e8473a462981149bb03db99f22d11e61f85c397dbf0: openshift-multus/multus-additional-cni-plugins-5fgxc/egress-router-binary-copy" id=bd32f92b-c3d3-46e3-a46d-ec87e47354a7 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:40.904762547Z" level=info msg="Starting container: ffe41dec05c7f28d2bbc3e8473a462981149bb03db99f22d11e61f85c397dbf0" id=43df1c1d-5a86-4841-8588-3925914fcc32 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:46:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:40.974145913Z" level=info msg="Started container" PID=2732 containerID=d628617bbda48d44a43a225bf2f5e855121106f3afd330a3fdcb502bb0180e53 description=openshift-multus/multus-4xhw7/kube-multus id=91a9e4f8-bcd8-4039-9807-463f572690b5 name=/runtime.v1.RuntimeService/StartContainer sandboxID=c2921d3b0822033449151099f7e4a989b1b0cb7526ca50c89a34c94fc6dd7157 Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.009514164Z" level=info msg="Created container 891a7a879802bf5ea7cd6f6f72e0369820a99e3061527667ecff8c5763f493d0: openshift-image-registry/node-ca-xw6d2/node-ca" id=0d336939-5913-4e93-a8aa-72ddc80e1e82 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.010036291Z" level=info msg="Starting container: 891a7a879802bf5ea7cd6f6f72e0369820a99e3061527667ecff8c5763f493d0" id=44d89590-b6da-4a02-b8ea-92be1b76e81d name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.030058474Z" level=info msg="Created container b7c254835fed9e60aa6d905f99f151ade924d428710ed31cdc2c8e5676bed6e5: openshift-dns/node-resolver-nxzr8/dns-node-resolver" id=3758c9ee-091d-4401-abb6-41c15bc25d44 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.030475025Z" level=info msg="Starting container: b7c254835fed9e60aa6d905f99f151ade924d428710ed31cdc2c8e5676bed6e5" id=40c84994-2729-4a2a-8b73-ed70a50bd375 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.057038466Z" level=info msg="Started container" PID=2706 containerID=ffe41dec05c7f28d2bbc3e8473a462981149bb03db99f22d11e61f85c397dbf0 description=openshift-multus/multus-additional-cni-plugins-5fgxc/egress-router-binary-copy id=43df1c1d-5a86-4841-8588-3925914fcc32 name=/runtime.v1.RuntimeService/StartContainer sandboxID=0acf147f6c1cead086fc9a6bc2a1864cb87fecc6e2901c79505a3a3c2d4136d8 Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.077611069Z" level=info msg="Created container 195db79f61f2dc0b97a2d254c01a536c649077f6a127e7a1b4ab858c1fee3a33: openshift-monitoring/node-exporter-2r6nf/node-exporter" id=a78e8163-64b1-4755-b378-35ae8d60bba0 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.078345835Z" level=info msg="Starting container: 195db79f61f2dc0b97a2d254c01a536c649077f6a127e7a1b4ab858c1fee3a33" id=e1d31848-4538-4373-bc35-afff2de51da0 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.092902308Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_55ca5731-353f-41ca-86f6-098734730b66\"" Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.107494987Z" level=info msg="Started container" PID=2808 containerID=b7c254835fed9e60aa6d905f99f151ade924d428710ed31cdc2c8e5676bed6e5 description=openshift-dns/node-resolver-nxzr8/dns-node-resolver id=40c84994-2729-4a2a-8b73-ed70a50bd375 name=/runtime.v1.RuntimeService/StartContainer sandboxID=09aeccea65d9703405b33f05b7cad2a5bb0bc55a0cbc0c82b5c23b0e4272ecaa Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.153602258Z" level=info msg="Started container" PID=2799 containerID=891a7a879802bf5ea7cd6f6f72e0369820a99e3061527667ecff8c5763f493d0 description=openshift-image-registry/node-ca-xw6d2/node-ca id=44d89590-b6da-4a02-b8ea-92be1b76e81d name=/runtime.v1.RuntimeService/StartContainer sandboxID=66a75af31e155b3ae3bef9faad30f7bd130b43e5719f5074ae6205394f8db4f7 Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.204864938Z" level=info msg="Started container" PID=2833 containerID=195db79f61f2dc0b97a2d254c01a536c649077f6a127e7a1b4ab858c1fee3a33 description=openshift-monitoring/node-exporter-2r6nf/node-exporter id=e1d31848-4538-4373-bc35-afff2de51da0 name=/runtime.v1.RuntimeService/StartContainer sandboxID=bccfa67f1ae302d09897f752660b0b20a0bf1f6fa4b4d3a20946611940b2e44e Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.260487716Z" level=info msg="Created container 514c61a70e1a00ce268eb8b5ca43491aabb497e894febead79e6d9d2b7f68b12: openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-liveness-probe" id=978b41d1-3413-4f50-a5d1-ada60633b9c4 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.260990169Z" level=info msg="Starting container: 514c61a70e1a00ce268eb8b5ca43491aabb497e894febead79e6d9d2b7f68b12" id=1562b533-d2ed-42c5-acad-faeba3719b28 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.268124244Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=a08440f5-effa-4ced-b8d7-9a2fd5eb9ca1 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.268345319Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=a08440f5-effa-4ced-b8d7-9a2fd5eb9ca1 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.271380477Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=00be6fe4-95aa-4526-9854-d2edcfc84b83 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.271588569Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=00be6fe4-95aa-4526-9854-d2edcfc84b83 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.273010022Z" level=info msg="Creating container: openshift-monitoring/node-exporter-2r6nf/kube-rbac-proxy" id=2923c2b1-8247-4f7d-9f04-6adfd37c6d96 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.273139125Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.283869881Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.283893018Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.283910039Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_c66f53b2-c5d6-4c23-af9a-4d2b21b1eac3\"" Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.322281641Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.322313391Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.324497363Z" level=info msg="Started container" PID=2927 containerID=514c61a70e1a00ce268eb8b5ca43491aabb497e894febead79e6d9d2b7f68b12 description=openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4/csi-liveness-probe id=1562b533-d2ed-42c5-acad-faeba3719b28 name=/runtime.v1.RuntimeService/StartContainer sandboxID=d7e9f4289f7cd500918dbb78134ef857afc6259662f81e9e13bb19cba691b342 Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:41.452450 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-hf8f5" event=&{ID:997b7ab4-939e-465c-9c7d-4a2ebe3a797a Type:ContainerStarted Data:db2c1a2ca7fd456ff77bdd0c3b85bc2af46b67cf77238b4d8471a3e5d72799d6} Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:41.454680 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4xhw7" event=&{ID:84f9cd8a-5e58-4aff-aedd-906ed37dc97d Type:ContainerStarted Data:d628617bbda48d44a43a225bf2f5e855121106f3afd330a3fdcb502bb0180e53} Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:41.463967 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-nxzr8" event=&{ID:8f88cbe5-1221-4e6f-a6c9-f22da152b43f Type:ContainerStarted Data:b7c254835fed9e60aa6d905f99f151ade924d428710ed31cdc2c8e5676bed6e5} Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:41.464091 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerStarted Data:ffe41dec05c7f28d2bbc3e8473a462981149bb03db99f22d11e61f85c397dbf0} Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:41.470159 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-2r6nf" event=&{ID:aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f Type:ContainerStarted Data:195db79f61f2dc0b97a2d254c01a536c649077f6a127e7a1b4ab858c1fee3a33} Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:41.476568 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" event=&{ID:1bb84a70-232c-4467-ac0b-647030155908 Type:ContainerStarted Data:514c61a70e1a00ce268eb8b5ca43491aabb497e894febead79e6d9d2b7f68b12} Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:41.476597 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-csi-drivers/alibaba-disk-csi-driver-node-5sqb4" event=&{ID:1bb84a70-232c-4467-ac0b-647030155908 Type:ContainerStarted Data:4ea62ec0ac85653a3c58bb51ccde4d6829ab49156353196af96d1e63801074da} Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:41.479023 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-sdn/sdn-66nsp" event=&{ID:d5b0f5e6-3d70-4315-b9c5-1403036f517b Type:ContainerStarted Data:ce8da87ec13d04b847764bc661e8abc9371a4ebe65e568a5847aadad12983ad3} Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:41.480737 1977 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-sdn/sdn-66nsp" Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:41.482212 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-xw6d2" event=&{ID:659b8bb4-81b5-49bf-8127-1d326f30ba2f Type:ContainerStarted Data:891a7a879802bf5ea7cd6f6f72e0369820a99e3061527667ecff8c5763f493d0} Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.503635060Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/multus\"" Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.570452994Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.570819271Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.570989352Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/egress-router\"" Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.600968572Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.600998358Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.601018565Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_55ca5731-353f-41ca-86f6-098734730b66\"" Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.619196494Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_c66f53b2-c5d6-4c23-af9a-4d2b21b1eac3\"" Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.828967090Z" level=info msg="Created container c9a42869a588ac99ff825373d7d8df296a0e78ab733de8a573316c895613b346: openshift-monitoring/node-exporter-2r6nf/kube-rbac-proxy" id=2923c2b1-8247-4f7d-9f04-6adfd37c6d96 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.829544650Z" level=info msg="Starting container: c9a42869a588ac99ff825373d7d8df296a0e78ab733de8a573316c895613b346" id=16515e61-812c-4fcc-af02-bf9971f23995 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:46:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:41.849244072Z" level=info msg="Started container" PID=3048 containerID=c9a42869a588ac99ff825373d7d8df296a0e78ab733de8a573316c895613b346 description=openshift-monitoring/node-exporter-2r6nf/kube-rbac-proxy id=16515e61-812c-4fcc-af02-bf9971f23995 name=/runtime.v1.RuntimeService/StartContainer sandboxID=bccfa67f1ae302d09897f752660b0b20a0bf1f6fa4b4d3a20946611940b2e44e Jan 05 09:46:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:42.264646646Z" level=info msg="CNI monitoring event REMOVE \"/etc/kubernetes/cni/net.d/00-multus.conf\"" Jan 05 09:46:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:42.265163696Z" level=info msg="Updated default CNI network name to " Jan 05 09:46:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:42.265323040Z" level=info msg="CNI monitoring event CREATE \"/etc/kubernetes/cni/net.d/00-multus.conf\"" Jan 05 09:46:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:42.265483555Z" level=error msg="Error loading CNI config file /etc/kubernetes/cni/net.d/00-multus.conf: error parsing configuration: unexpected end of JSON input" Jan 05 09:46:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:42.265508385Z" level=info msg="Updated default CNI network name to " Jan 05 09:46:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:42.265592732Z" level=info msg="CNI monitoring event WRITE \"/etc/kubernetes/cni/net.d/00-multus.conf\"" Jan 05 09:46:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:42.285722405Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:46:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:42.285754031Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:46:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:42.285769195Z" level=info msg="CNI monitoring event CHMOD \"/etc/kubernetes/cni/net.d/00-multus.conf\"" Jan 05 09:46:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:42.487724 1977 generic.go:296] "Generic (PLEG): container finished" podID=c0e47249-a5ff-460e-958d-53577387c154 containerID="ffe41dec05c7f28d2bbc3e8473a462981149bb03db99f22d11e61f85c397dbf0" exitCode=0 Jan 05 09:46:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:42.489913044Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d5fc161170fcf68900b2694039803a2e7f1b31eb913566139a41f91851f166d7" id=11f76c26-9a69-4faa-83c8-001e54580e1a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:42.490286764Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:852cc803756e640141d6c3eb3e6607f2b0f7076ea77f9190b1f490288f40b3d8,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d5fc161170fcf68900b2694039803a2e7f1b31eb913566139a41f91851f166d7],Size_:574516705,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=11f76c26-9a69-4faa-83c8-001e54580e1a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:42.487880 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerDied Data:ffe41dec05c7f28d2bbc3e8473a462981149bb03db99f22d11e61f85c397dbf0} Jan 05 09:46:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:42.497423 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-2r6nf" event=&{ID:aa84cbad-3d0f-43bd-a8a9-46dc1cd9c57f Type:ContainerStarted Data:c9a42869a588ac99ff825373d7d8df296a0e78ab733de8a573316c895613b346} Jan 05 09:46:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:42.502001899Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d5fc161170fcf68900b2694039803a2e7f1b31eb913566139a41f91851f166d7" id=0a5d73c1-c19e-464d-a93d-3aa95a67702e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:42.502371732Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:852cc803756e640141d6c3eb3e6607f2b0f7076ea77f9190b1f490288f40b3d8,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:d5fc161170fcf68900b2694039803a2e7f1b31eb913566139a41f91851f166d7],Size_:574516705,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=0a5d73c1-c19e-464d-a93d-3aa95a67702e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:42.506974630Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/cni-plugins" id=63e270c3-4137-41fc-96a2-4b966a7ff7cf name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:42.507462237Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:46:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: 2023-01-05T09:46:42Z [verbose] Del: openshift-dns:dns-default-2pfzf:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:46:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:42.895172050Z" level=info msg="Successfully cleaned up network for pod 4f0a921bd0d3c8f309751665c310e818034ecd3252406f2a101ca811d04377c5" Jan 05 09:46:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:42.895210623Z" level=info msg="cleanup sandbox network" Jan 05 09:46:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:42.895222384Z" level=info msg="Successfully cleaned up network for pod e0df29cbf3197efe5e4c221145d0f72e155c977f3d50c18c2f38862f24a5522e" Jan 05 09:46:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:42.895233204Z" level=info msg="cleanup sandbox network" Jan 05 09:46:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:42.895243111Z" level=info msg="Successfully cleaned up network for pod acd08c77f0076092e020bedcc20e2c94faab21df9b7d5516eba6f0dbe370935b" Jan 05 09:46:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:42.895253610Z" level=info msg="cleanup sandbox network" Jan 05 09:46:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:42.897000474Z" level=info msg="Got pod network &{Name:dns-default-2pfzf Namespace:openshift-dns ID:d5240e19e88cdeb551996f336d62881fb4970c3105fcfbd0a1f4ba0ff48acf04 UID:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 NetNS:/var/run/netns/fb949b20-a423-40fd-9c48-28f2f8e6d9d2 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:46:42 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:42.897195833Z" level=info msg="Adding pod openshift-dns_dns-default-2pfzf to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.113672976Z" level=info msg="Created container 2b57f408a20062c96d24c38b0d5c8b7314a34fd56953d48903c7a4391209b952: openshift-multus/multus-additional-cni-plugins-5fgxc/cni-plugins" id=63e270c3-4137-41fc-96a2-4b966a7ff7cf name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.114697589Z" level=info msg="Starting container: 2b57f408a20062c96d24c38b0d5c8b7314a34fd56953d48903c7a4391209b952" id=84c90bd7-58d2-4458-b5e3-b5a4934f8ebb name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.136089799Z" level=info msg="Started container" PID=3463 containerID=2b57f408a20062c96d24c38b0d5c8b7314a34fd56953d48903c7a4391209b952 description=openshift-multus/multus-additional-cni-plugins-5fgxc/cni-plugins id=84c90bd7-58d2-4458-b5e3-b5a4934f8ebb name=/runtime.v1.RuntimeService/StartContainer sandboxID=0acf147f6c1cead086fc9a6bc2a1864cb87fecc6e2901c79505a3a3c2d4136d8 Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.170825301Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_dcf7c31b-caee-444a-bf4c-4ebed3f34d82\"" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.188289488Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.188322501Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.362656482Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/bandwidth\"" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.374261013Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.374435482Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.374544575Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/bridge\"" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.398514070Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.398548745Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.398566059Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/dhcp\"" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.414097362Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.414124923Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.414141151Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/firewall\"" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.430907599Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.430937497Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.430969425Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/host-device\"" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.448363201Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.448388091Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.448404433Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/host-local\"" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.472487642Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.472513946Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.472532813Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/ipvlan\"" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.499344022Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.499369797Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.499386146Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/loopback\"" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:43.503022 1977 generic.go:296] "Generic (PLEG): container finished" podID=c0e47249-a5ff-460e-958d-53577387c154 containerID="2b57f408a20062c96d24c38b0d5c8b7314a34fd56953d48903c7a4391209b952" exitCode=0 Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:43.503595 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerDied Data:2b57f408a20062c96d24c38b0d5c8b7314a34fd56953d48903c7a4391209b952} Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.507440007Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7ad3b9824ea195f38b6346262f0c609415c487ffaeb554725621a8327266da0f" id=ceedc62d-5ec9-490c-9137-1a3c0ebed84e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.507664032Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:4c7cce1e79d06b3bc147fc59bd0a72991be97e767193f661db48974f3bb9ac30,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7ad3b9824ea195f38b6346262f0c609415c487ffaeb554725621a8327266da0f],Size_:336119748,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=ceedc62d-5ec9-490c-9137-1a3c0ebed84e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.508517123Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7ad3b9824ea195f38b6346262f0c609415c487ffaeb554725621a8327266da0f" id=849ffb5b-e83e-4fa4-abfd-27af042f9158 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.508716710Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:4c7cce1e79d06b3bc147fc59bd0a72991be97e767193f661db48974f3bb9ac30,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7ad3b9824ea195f38b6346262f0c609415c487ffaeb554725621a8327266da0f],Size_:336119748,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=849ffb5b-e83e-4fa4-abfd-27af042f9158 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.509912519Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/bond-cni-plugin" id=8409cb5e-785a-4101-8433-8c128eac43a0 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.510046473Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.515001162Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.515101208Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.515177318Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/macvlan\"" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.526902702Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.526924638Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.526939864Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/portmap\"" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.545108352Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.545129911Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.545144384Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/ptp\"" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.555864988Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.555883390Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.555897369Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/sbr\"" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.564495660Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.564515242Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.564529070Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/static\"" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.577052540Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.577070510Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.577084021Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/tuning\"" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.587023640Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.587043062Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.587058626Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/vlan\"" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.595469087Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.595486665Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.595500110Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/vrf\"" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.606782360Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.606800500Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.606814801Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_dcf7c31b-caee-444a-bf4c-4ebed3f34d82\"" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:43.792200 1977 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-network-diagnostics/network-check-target-xmq2g] Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: 2023-01-05T09:46:43Z [verbose] Add: openshift-network-diagnostics:network-check-target-xmq2g:f364a949-09d7-4cfb-83ff-e532b822a557:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/53f617ed-730b-4fb4-abb1-c8d0d0bdef7e"}],"ips":[{"version":"4","interface":0,"address":"10.131.0.3/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: I0105 09:46:43.772020 2118 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-network-diagnostics", Name:"network-check-target-xmq2g", UID:"f364a949-09d7-4cfb-83ff-e532b822a557", APIVersion:"v1", ResourceVersion:"107204", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.3/23] from openshift-sdn Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.802698471Z" level=info msg="Got pod network &{Name:network-check-target-xmq2g Namespace:openshift-network-diagnostics ID:214a9c55e49c804acc78be4ce16b5a653ad517d1495dac27266eba0b4f45de68 UID:f364a949-09d7-4cfb-83ff-e532b822a557 NetNS:/var/run/netns/53f617ed-730b-4fb4-abb1-c8d0d0bdef7e Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.803277903Z" level=info msg="Checking pod openshift-network-diagnostics_network-check-target-xmq2g for CNI network multus-cni-network (type=multus)" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: W0105 09:46:43.813583 1977 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf364a949_09d7_4cfb_83ff_e532b822a557.slice/crio-214a9c55e49c804acc78be4ce16b5a653ad517d1495dac27266eba0b4f45de68.scope WatchSource:0}: Error finding container 214a9c55e49c804acc78be4ce16b5a653ad517d1495dac27266eba0b4f45de68: Status 404 returned error can't find the container with id 214a9c55e49c804acc78be4ce16b5a653ad517d1495dac27266eba0b4f45de68 Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.820261733Z" level=info msg="Ran pod sandbox 214a9c55e49c804acc78be4ce16b5a653ad517d1495dac27266eba0b4f45de68 with infra container: openshift-network-diagnostics/network-check-target-xmq2g/POD" id=84839fef-8cc1-44ae-9dfd-90086ea73636 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.836258159Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7" id=49846585-60aa-42be-9b40-73b40ddbc52f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.836611789Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:4cb41c73a232ec4401c5db32cdb523810dc12ac605cdc6a9ce388e90930eb6c4,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7],Size_:515955107,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=49846585-60aa-42be-9b40-73b40ddbc52f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.842390940Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7" id=fbd239c2-7211-4a1b-a42a-7999807a0ac7 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.842719074Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:4cb41c73a232ec4401c5db32cdb523810dc12ac605cdc6a9ce388e90930eb6c4,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7],Size_:515955107,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=fbd239c2-7211-4a1b-a42a-7999807a0ac7 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.846091897Z" level=info msg="Creating container: openshift-network-diagnostics/network-check-target-xmq2g/network-check-target-container" id=cc906ba6-bbb2-4e69-8efd-3650730df670 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.846233360Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.927866474Z" level=info msg="Created container 7ae35548e5d9248afea3f2522362448c992df59de3e014bfe95a1b6f5105211d: openshift-multus/multus-additional-cni-plugins-5fgxc/bond-cni-plugin" id=8409cb5e-785a-4101-8433-8c128eac43a0 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.928522283Z" level=info msg="Starting container: 7ae35548e5d9248afea3f2522362448c992df59de3e014bfe95a1b6f5105211d" id=66412448-6e71-4823-9adf-355a52f8cbbb name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.944416248Z" level=info msg="Started container" PID=3650 containerID=7ae35548e5d9248afea3f2522362448c992df59de3e014bfe95a1b6f5105211d description=openshift-multus/multus-additional-cni-plugins-5fgxc/bond-cni-plugin id=66412448-6e71-4823-9adf-355a52f8cbbb name=/runtime.v1.RuntimeService/StartContainer sandboxID=0acf147f6c1cead086fc9a6bc2a1864cb87fecc6e2901c79505a3a3c2d4136d8 Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.965540138Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_ab8397d0-b6a0-427c-a975-8108e982b9bc\"" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.986977901Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.987108441Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:46:43 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:43.987192032Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/bond\"" Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.032008558Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.032041769Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.032061069Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_ab8397d0-b6a0-427c-a975-8108e982b9bc\"" Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.049660581Z" level=info msg="Created container d2116d102724d91970c56a20fbcf7abfc4e11bbd00d6d96861e2a1b03a66bfc6: openshift-network-diagnostics/network-check-target-xmq2g/network-check-target-container" id=cc906ba6-bbb2-4e69-8efd-3650730df670 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.050460252Z" level=info msg="Starting container: d2116d102724d91970c56a20fbcf7abfc4e11bbd00d6d96861e2a1b03a66bfc6" id=f4077877-bd81-4149-9aac-c9558ada1a8d name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.061319921Z" level=info msg="Started container" PID=3671 containerID=d2116d102724d91970c56a20fbcf7abfc4e11bbd00d6d96861e2a1b03a66bfc6 description=openshift-network-diagnostics/network-check-target-xmq2g/network-check-target-container id=f4077877-bd81-4149-9aac-c9558ada1a8d name=/runtime.v1.RuntimeService/StartContainer sandboxID=214a9c55e49c804acc78be4ce16b5a653ad517d1495dac27266eba0b4f45de68 Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:44.069270 1977 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-multus/network-metrics-daemon-hq594] Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: 2023-01-05T09:46:44Z [verbose] Add: openshift-multus:network-metrics-daemon-hq594:50152296-9489-4fa8-aa42-7858debe1c08:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/ba9d8790-155d-4a0b-bb29-8fbef77d7fe5"}],"ips":[{"version":"4","interface":0,"address":"10.131.0.4/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: I0105 09:46:44.055018 2141 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-multus", Name:"network-metrics-daemon-hq594", UID:"50152296-9489-4fa8-aa42-7858debe1c08", APIVersion:"v1", ResourceVersion:"107210", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.4/23] from openshift-sdn Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.072092932Z" level=info msg="Got pod network &{Name:network-metrics-daemon-hq594 Namespace:openshift-multus ID:7e8ab5f06c8a76365c09171cecfc24f31a6a7f245ad42d5208ec29334d98b2c0 UID:50152296-9489-4fa8-aa42-7858debe1c08 NetNS:/var/run/netns/ba9d8790-155d-4a0b-bb29-8fbef77d7fe5 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.072314371Z" level=info msg="Checking pod openshift-multus_network-metrics-daemon-hq594 for CNI network multus-cni-network (type=multus)" Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: W0105 09:46:44.080334 1977 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50152296_9489_4fa8_aa42_7858debe1c08.slice/crio-7e8ab5f06c8a76365c09171cecfc24f31a6a7f245ad42d5208ec29334d98b2c0.scope WatchSource:0}: Error finding container 7e8ab5f06c8a76365c09171cecfc24f31a6a7f245ad42d5208ec29334d98b2c0: Status 404 returned error can't find the container with id 7e8ab5f06c8a76365c09171cecfc24f31a6a7f245ad42d5208ec29334d98b2c0 Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.082915664Z" level=info msg="Ran pod sandbox 7e8ab5f06c8a76365c09171cecfc24f31a6a7f245ad42d5208ec29334d98b2c0 with infra container: openshift-multus/network-metrics-daemon-hq594/POD" id=3b85e0f9-6bf9-45bc-b186-ac78e116f33e name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.084181413Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9e33552d1aaea9cf6a301fa438cfc8305826c5ddb58ac8437d884e55e71c6e44" id=09e4b403-fefc-4941-b528-84258dbf1699 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.084380192Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:7b87fc786118817b8913f28a61e01db37b08380f754bbaacf425508b208632a0,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9e33552d1aaea9cf6a301fa438cfc8305826c5ddb58ac8437d884e55e71c6e44],Size_:312894682,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=09e4b403-fefc-4941-b528-84258dbf1699 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.085977721Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9e33552d1aaea9cf6a301fa438cfc8305826c5ddb58ac8437d884e55e71c6e44" id=1350a517-89a3-4aab-8525-f90971455edd name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.086246457Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:7b87fc786118817b8913f28a61e01db37b08380f754bbaacf425508b208632a0,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9e33552d1aaea9cf6a301fa438cfc8305826c5ddb58ac8437d884e55e71c6e44],Size_:312894682,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=1350a517-89a3-4aab-8525-f90971455edd name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.086735473Z" level=info msg="Creating container: openshift-multus/network-metrics-daemon-hq594/network-metrics-daemon" id=d60175a6-6969-4522-b426-1584e1ab8592 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.086830460Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:44.239803 1977 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-dns/dns-default-2pfzf] Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: 2023-01-05T09:46:44Z [verbose] Add: openshift-dns:dns-default-2pfzf:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/fb949b20-a423-40fd-9c48-28f2f8e6d9d2"}],"ips":[{"version":"4","interface":0,"address":"10.131.0.5/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: I0105 09:46:44.204983 3447 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-dns", Name:"dns-default-2pfzf", UID:"1cc2aadb-14f2-4d6e-97fc-6fdab8889a63", APIVersion:"v1", ResourceVersion:"107224", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.5/23] from openshift-sdn Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.240637572Z" level=info msg="Got pod network &{Name:dns-default-2pfzf Namespace:openshift-dns ID:d5240e19e88cdeb551996f336d62881fb4970c3105fcfbd0a1f4ba0ff48acf04 UID:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 NetNS:/var/run/netns/fb949b20-a423-40fd-9c48-28f2f8e6d9d2 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.240775354Z" level=info msg="Checking pod openshift-dns_dns-default-2pfzf for CNI network multus-cni-network (type=multus)" Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: W0105 09:46:44.249142 1977 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1cc2aadb_14f2_4d6e_97fc_6fdab8889a63.slice/crio-d5240e19e88cdeb551996f336d62881fb4970c3105fcfbd0a1f4ba0ff48acf04.scope WatchSource:0}: Error finding container d5240e19e88cdeb551996f336d62881fb4970c3105fcfbd0a1f4ba0ff48acf04: Status 404 returned error can't find the container with id d5240e19e88cdeb551996f336d62881fb4970c3105fcfbd0a1f4ba0ff48acf04 Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.251389325Z" level=info msg="Ran pod sandbox d5240e19e88cdeb551996f336d62881fb4970c3105fcfbd0a1f4ba0ff48acf04 with infra container: openshift-dns/dns-default-2pfzf/POD" id=7a11fa5d-3411-41a3-a3fe-df3e594289bf name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.252229458Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f674b597694b13b7060c9ee611a403f1f87a98e05682c33f9a86140061ee9d9e" id=55fe95da-4afa-4245-a180-371e81a3f327 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.252476414Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:cdb247afcbd9968f1a59680dcc4de0364f7ddd8fc25dd3a12db59f05c8e026a0,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f674b597694b13b7060c9ee611a403f1f87a98e05682c33f9a86140061ee9d9e],Size_:404212434,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=55fe95da-4afa-4245-a180-371e81a3f327 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.253164921Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f674b597694b13b7060c9ee611a403f1f87a98e05682c33f9a86140061ee9d9e" id=8e9e29c8-993f-49ad-9897-20ab521a174f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.253334877Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:cdb247afcbd9968f1a59680dcc4de0364f7ddd8fc25dd3a12db59f05c8e026a0,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:f674b597694b13b7060c9ee611a403f1f87a98e05682c33f9a86140061ee9d9e],Size_:404212434,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=8e9e29c8-993f-49ad-9897-20ab521a174f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.254287605Z" level=info msg="Creating container: openshift-dns/dns-default-2pfzf/dns" id=35fded14-1816-40c0-9261-b35f4a4c1a68 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.254491191Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.261724484Z" level=info msg="Created container b02b52f0034e7b795430f3441f0bea920b4d1aafde89454f77440241edcf0c42: openshift-multus/network-metrics-daemon-hq594/network-metrics-daemon" id=d60175a6-6969-4522-b426-1584e1ab8592 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.262071359Z" level=info msg="Starting container: b02b52f0034e7b795430f3441f0bea920b4d1aafde89454f77440241edcf0c42" id=a53f04d1-5a4b-43a4-b17f-7716f6559e77 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.278069538Z" level=info msg="Started container" PID=3772 containerID=b02b52f0034e7b795430f3441f0bea920b4d1aafde89454f77440241edcf0c42 description=openshift-multus/network-metrics-daemon-hq594/network-metrics-daemon id=a53f04d1-5a4b-43a4-b17f-7716f6559e77 name=/runtime.v1.RuntimeService/StartContainer sandboxID=7e8ab5f06c8a76365c09171cecfc24f31a6a7f245ad42d5208ec29334d98b2c0 Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.291263124Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=27b75254-6a5f-4468-bf2d-d593461d39c4 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.291590903Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=27b75254-6a5f-4468-bf2d-d593461d39c4 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.292794888Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=4a5dd696-1983-404c-9095-7d11fd742578 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.292998477Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=4a5dd696-1983-404c-9095-7d11fd742578 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.293704014Z" level=info msg="Creating container: openshift-multus/network-metrics-daemon-hq594/kube-rbac-proxy" id=b95aab86-6ca6-4af7-ae08-1ee6ff216847 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.293919525Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:44.518304 1977 generic.go:296] "Generic (PLEG): container finished" podID=c0e47249-a5ff-460e-958d-53577387c154 containerID="7ae35548e5d9248afea3f2522362448c992df59de3e014bfe95a1b6f5105211d" exitCode=0 Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:44.518376 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerDied Data:7ae35548e5d9248afea3f2522362448c992df59de3e014bfe95a1b6f5105211d} Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.520648850Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:a8bcc8cf428e222c7e6677fe5f7f05e79da2aed1ded851cad696d83c94860b4c" id=cfa06bd7-d345-4079-9953-7f138c60cc3f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.521144945Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:34fa51fa29c1ebd2a3ad620255bb70edf89b0a96348f08efac3dc2ca3200602d,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:a8bcc8cf428e222c7e6677fe5f7f05e79da2aed1ded851cad696d83c94860b4c],Size_:275275974,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=cfa06bd7-d345-4079-9953-7f138c60cc3f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:44.527814 1977 plugin_watcher.go:203] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/diskplugin.csi.alibabacloud.com-reg.sock" Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.531147322Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:a8bcc8cf428e222c7e6677fe5f7f05e79da2aed1ded851cad696d83c94860b4c" id=ae502339-0b28-4a71-b7f3-dcbec8f6433c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.532754365Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:34fa51fa29c1ebd2a3ad620255bb70edf89b0a96348f08efac3dc2ca3200602d,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:a8bcc8cf428e222c7e6677fe5f7f05e79da2aed1ded851cad696d83c94860b4c],Size_:275275974,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=ae502339-0b28-4a71-b7f3-dcbec8f6433c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.535145109Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/routeoverride-cni" id=11cc189a-bea4-43a1-a9b3-23b53c26d24e name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.535352008Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:44.540054 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-2pfzf" event=&{ID:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 Type:ContainerStarted Data:d5240e19e88cdeb551996f336d62881fb4970c3105fcfbd0a1f4ba0ff48acf04} Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:44.561599 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xmq2g" event=&{ID:f364a949-09d7-4cfb-83ff-e532b822a557 Type:ContainerStarted Data:d2116d102724d91970c56a20fbcf7abfc4e11bbd00d6d96861e2a1b03a66bfc6} Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:44.561632 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xmq2g" event=&{ID:f364a949-09d7-4cfb-83ff-e532b822a557 Type:ContainerStarted Data:214a9c55e49c804acc78be4ce16b5a653ad517d1495dac27266eba0b4f45de68} Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:44.563892 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hq594" event=&{ID:50152296-9489-4fa8-aa42-7858debe1c08 Type:ContainerStarted Data:b02b52f0034e7b795430f3441f0bea920b4d1aafde89454f77440241edcf0c42} Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:44.563936 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hq594" event=&{ID:50152296-9489-4fa8-aa42-7858debe1c08 Type:ContainerStarted Data:7e8ab5f06c8a76365c09171cecfc24f31a6a7f245ad42d5208ec29334d98b2c0} Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:44.605657 1977 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-ingress-canary/ingress-canary-8xrbh] Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: 2023-01-05T09:46:44Z [verbose] Add: openshift-ingress-canary:ingress-canary-8xrbh:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/3f4b8877-65b1-4936-8a9d-2522145b5205"}],"ips":[{"version":"4","interface":0,"address":"10.131.0.2/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: I0105 09:46:44.553171 2107 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-ingress-canary", Name:"ingress-canary-8xrbh", UID:"e04a47b4-aa8e-4d4e-9d45-aa8037dcb748", APIVersion:"v1", ResourceVersion:"107218", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.2/23] from openshift-sdn Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.617610393Z" level=info msg="Got pod network &{Name:ingress-canary-8xrbh Namespace:openshift-ingress-canary ID:8510bb13dab2503578d194877e82efbb3c57ad0a439fe2f087985adeeca2f21b UID:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748 NetNS:/var/run/netns/3f4b8877-65b1-4936-8a9d-2522145b5205 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.618054643Z" level=info msg="Checking pod openshift-ingress-canary_ingress-canary-8xrbh for CNI network multus-cni-network (type=multus)" Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.619165814Z" level=info msg="Got pod network &{Name:ingress-canary-8xrbh Namespace:openshift-ingress-canary ID:2f42fdd8caa6c772e33c09307d855f2c6b126d16a962580bfad9af47fa893ef2 UID:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748 NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.619492300Z" level=info msg="Deleting pod openshift-ingress-canary_ingress-canary-8xrbh from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: W0105 09:46:44.634896 1977 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode04a47b4_aa8e_4d4e_9d45_aa8037dcb748.slice/crio-8510bb13dab2503578d194877e82efbb3c57ad0a439fe2f087985adeeca2f21b.scope WatchSource:0}: Error finding container 8510bb13dab2503578d194877e82efbb3c57ad0a439fe2f087985adeeca2f21b: Status 404 returned error can't find the container with id 8510bb13dab2503578d194877e82efbb3c57ad0a439fe2f087985adeeca2f21b Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.648753066Z" level=info msg="Ran pod sandbox 8510bb13dab2503578d194877e82efbb3c57ad0a439fe2f087985adeeca2f21b with infra container: openshift-ingress-canary/ingress-canary-8xrbh/POD" id=a7bbab8d-0492-4b98-8549-e10134404e6a name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.663532839Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:5974989cda2655deb3052d601e7406eaf3caee0e7999a6eff251760aac19c199" id=0d9c697f-f5a5-4cca-a138-40abbf5eea32 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.663856632Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:c34c97b229e46faa94bab6ad2559fd683bd036df06031c9097dd27f7daaf96c7,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:5974989cda2655deb3052d601e7406eaf3caee0e7999a6eff251760aac19c199],Size_:417780553,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=0d9c697f-f5a5-4cca-a138-40abbf5eea32 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.664807532Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:5974989cda2655deb3052d601e7406eaf3caee0e7999a6eff251760aac19c199" id=75bfab92-7955-4d21-8dc9-b50b0b081c27 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.665319906Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:c34c97b229e46faa94bab6ad2559fd683bd036df06031c9097dd27f7daaf96c7,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:5974989cda2655deb3052d601e7406eaf3caee0e7999a6eff251760aac19c199],Size_:417780553,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=75bfab92-7955-4d21-8dc9-b50b0b081c27 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.667184076Z" level=info msg="Creating container: openshift-ingress-canary/ingress-canary-8xrbh/serve-healthcheck-canary" id=8b61ac80-dd28-43bf-88ab-ce119fa339db name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.667414371Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.737048736Z" level=info msg="Created container 68768e66c297fd1239a4bcc9c0fadad2e81bda22e11b25f192f1680d34520a8a: openshift-multus/network-metrics-daemon-hq594/kube-rbac-proxy" id=b95aab86-6ca6-4af7-ae08-1ee6ff216847 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.740767532Z" level=info msg="Starting container: 68768e66c297fd1239a4bcc9c0fadad2e81bda22e11b25f192f1680d34520a8a" id=a0f60e45-435b-408a-bce8-9a02de33aabb name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.767069396Z" level=info msg="Started container" PID=3834 containerID=68768e66c297fd1239a4bcc9c0fadad2e81bda22e11b25f192f1680d34520a8a description=openshift-multus/network-metrics-daemon-hq594/kube-rbac-proxy id=a0f60e45-435b-408a-bce8-9a02de33aabb name=/runtime.v1.RuntimeService/StartContainer sandboxID=7e8ab5f06c8a76365c09171cecfc24f31a6a7f245ad42d5208ec29334d98b2c0 Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.833333589Z" level=info msg="Created container 102ae4f3b08fda6cde21713ce3ea0aa9d714ac1683607033e84094ae48a306c4: openshift-dns/dns-default-2pfzf/dns" id=35fded14-1816-40c0-9261-b35f4a4c1a68 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.833941658Z" level=info msg="Starting container: 102ae4f3b08fda6cde21713ce3ea0aa9d714ac1683607033e84094ae48a306c4" id=d91491c8-dbd0-46b5-a6d0-f9522d49a3bc name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.871032885Z" level=info msg="Started container" PID=3864 containerID=102ae4f3b08fda6cde21713ce3ea0aa9d714ac1683607033e84094ae48a306c4 description=openshift-dns/dns-default-2pfzf/dns id=d91491c8-dbd0-46b5-a6d0-f9522d49a3bc name=/runtime.v1.RuntimeService/StartContainer sandboxID=d5240e19e88cdeb551996f336d62881fb4970c3105fcfbd0a1f4ba0ff48acf04 Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: 2023-01-05T09:46:44Z [verbose] Del: openshift-ingress-canary:ingress-canary-8xrbh:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.884230333Z" level=info msg="Successfully cleaned up network for pod 2f42fdd8caa6c772e33c09307d855f2c6b126d16a962580bfad9af47fa893ef2" Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.884254826Z" level=info msg="cleanup sandbox network" Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.884265237Z" level=info msg="Successfully cleaned up network for pod 93662ac8682d21720eb888cc865f999977edd4dcd1e585e0bae6c3aeeb2d9bba" Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.884279421Z" level=info msg="cleanup sandbox network" Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.884938340Z" level=info msg="Got pod network &{Name:network-metrics-daemon-hq594 Namespace:openshift-multus ID:44ef9ac88935fc68d10995ab1204d3f92afbedeae9fabaf8c0abb737181fa9fd UID:50152296-9489-4fa8-aa42-7858debe1c08 NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.885108162Z" level=info msg="Deleting pod openshift-multus_network-metrics-daemon-hq594 from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.914090141Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=23cbcadf-d159-4fd9-b7b4-f0ea50533635 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.914482133Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=23cbcadf-d159-4fd9-b7b4-f0ea50533635 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.915479067Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=04b61042-73e5-4910-8880-606b2e1ca5c2 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.915673252Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=04b61042-73e5-4910-8880-606b2e1ca5c2 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.919665551Z" level=info msg="Creating container: openshift-dns/dns-default-2pfzf/kube-rbac-proxy" id=807f3b04-8b9f-4e1a-8f3e-1719693e05cd name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:44 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:44.919818185Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:45.008514 1977 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: 2023-01-05T09:46:45Z [verbose] Del: openshift-multus:network-metrics-daemon-hq594:50152296-9489-4fa8-aa42-7858debe1c08:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:45.153170710Z" level=info msg="Successfully cleaned up network for pod 44ef9ac88935fc68d10995ab1204d3f92afbedeae9fabaf8c0abb737181fa9fd" Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:45.153202214Z" level=info msg="cleanup sandbox network" Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:45.153761576Z" level=info msg="Got pod network &{Name:network-check-target-xmq2g Namespace:openshift-network-diagnostics ID:df03e4f874ec90019c812fc416c2d12340d64e387a2a542bdc3beee879bc31bd UID:f364a949-09d7-4cfb-83ff-e532b822a557 NetNS: Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:45.153933432Z" level=info msg="Deleting pod openshift-network-diagnostics_network-check-target-xmq2g from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:45.165372131Z" level=info msg="Created container da2062678f72ca3bca86691478453ec531e9372503c61018f96ef5bba86dbf23: openshift-ingress-canary/ingress-canary-8xrbh/serve-healthcheck-canary" id=8b61ac80-dd28-43bf-88ab-ce119fa339db name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:45.166065687Z" level=info msg="Starting container: da2062678f72ca3bca86691478453ec531e9372503c61018f96ef5bba86dbf23" id=2fbca67a-3899-4cbb-b40f-36463b87a50a name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:45.205348281Z" level=info msg="Started container" PID=3962 containerID=da2062678f72ca3bca86691478453ec531e9372503c61018f96ef5bba86dbf23 description=openshift-ingress-canary/ingress-canary-8xrbh/serve-healthcheck-canary id=2fbca67a-3899-4cbb-b40f-36463b87a50a name=/runtime.v1.RuntimeService/StartContainer sandboxID=8510bb13dab2503578d194877e82efbb3c57ad0a439fe2f087985adeeca2f21b Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:45.279536 1977 reconciler.go:164] "OperationExecutor.RegisterPlugin started" plugin={SocketPath:/var/lib/kubelet/plugins_registry/diskplugin.csi.alibabacloud.com-reg.sock Timestamp:2023-01-05 09:46:44.527833001 +0000 UTC m=+7.963455494 Handler: Name:} Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:45.295330 1977 csi_plugin.go:99] kubernetes.io/csi: Trying to validate a new CSI Driver with name: diskplugin.csi.alibabacloud.com endpoint: /var/lib/kubelet/plugins/diskplugin.csi.alibabacloud.com/csi.sock versions: 1.0.0 Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:45.295360 1977 csi_plugin.go:112] kubernetes.io/csi: Register new plugin with name: diskplugin.csi.alibabacloud.com at endpoint: /var/lib/kubelet/plugins/diskplugin.csi.alibabacloud.com/csi.sock Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: 2023-01-05T09:46:45Z [verbose] Del: openshift-network-diagnostics:network-check-target-xmq2g:f364a949-09d7-4cfb-83ff-e532b822a557:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:45.329050702Z" level=info msg="Successfully cleaned up network for pod df03e4f874ec90019c812fc416c2d12340d64e387a2a542bdc3beee879bc31bd" Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:45.332724534Z" level=info msg="cleanup sandbox network" Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:45.334315232Z" level=info msg="Successfully cleaned up network for pod 91eb7fe6d0c596f220833b072fadd2715083a9931fdb3ed2bbf7986380dce490" Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:45.334420185Z" level=info msg="cleanup sandbox network" Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:45.334487673Z" level=info msg="Successfully cleaned up network for pod 16cbcad68c73abf56080eb14225620b7f239901aad0b8916effb3d6a9dc1a771" Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:45.334505294Z" level=info msg="cleanup sandbox network" Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:45.334514478Z" level=info msg="Successfully cleaned up network for pod bb5b6562ae64906e3450c53a90bf4df6ea55af8ce39868944ba2cdcf7eab1188" Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:45.334525056Z" level=info msg="cleanup sandbox network" Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:45.334534868Z" level=info msg="Successfully cleaned up network for pod ae18be6d794af60a14ce35ff3392698b0b7e426a5128047f31455ade1f2b79dd" Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:45.334545250Z" level=info msg="cleanup sandbox network" Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:45.334554738Z" level=info msg="Successfully cleaned up network for pod 06a6d838dd853ce187ccf364676fd4e36976cae02f20a98aeda4b59df6df2d93" Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:45.431803683Z" level=info msg="Created container 0b790541b89bed33d7866e16060250fae888afd249ff6773c7e8c5298843b39f: openshift-dns/dns-default-2pfzf/kube-rbac-proxy" id=807f3b04-8b9f-4e1a-8f3e-1719693e05cd name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:45.434746488Z" level=info msg="Starting container: 0b790541b89bed33d7866e16060250fae888afd249ff6773c7e8c5298843b39f" id=70fef566-e561-42cb-9c3e-3e0eee6b6107 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:45.456371338Z" level=info msg="Started container" PID=4047 containerID=0b790541b89bed33d7866e16060250fae888afd249ff6773c7e8c5298843b39f description=openshift-dns/dns-default-2pfzf/kube-rbac-proxy id=70fef566-e561-42cb-9c3e-3e0eee6b6107 name=/runtime.v1.RuntimeService/StartContainer sandboxID=d5240e19e88cdeb551996f336d62881fb4970c3105fcfbd0a1f4ba0ff48acf04 Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:45.491937089Z" level=info msg="Created container 936f593239e32a73c99f411c5dd4f6c509a35e0bc0c15d9738ea4de3b5de496a: openshift-multus/multus-additional-cni-plugins-5fgxc/routeoverride-cni" id=11cc189a-bea4-43a1-a9b3-23b53c26d24e name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:45.492355837Z" level=info msg="Starting container: 936f593239e32a73c99f411c5dd4f6c509a35e0bc0c15d9738ea4de3b5de496a" id=64778466-7644-42ad-8bc0-be1ba161bf3e name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:45.505360995Z" level=info msg="Started container" PID=4066 containerID=936f593239e32a73c99f411c5dd4f6c509a35e0bc0c15d9738ea4de3b5de496a description=openshift-multus/multus-additional-cni-plugins-5fgxc/routeoverride-cni id=64778466-7644-42ad-8bc0-be1ba161bf3e name=/runtime.v1.RuntimeService/StartContainer sandboxID=0acf147f6c1cead086fc9a6bc2a1864cb87fecc6e2901c79505a3a3c2d4136d8 Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:45.554427541Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_ec991d9d-3786-4a3f-962a-4dd7e96214da\"" Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:45.567589 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hq594" event=&{ID:50152296-9489-4fa8-aa42-7858debe1c08 Type:ContainerStarted Data:68768e66c297fd1239a4bcc9c0fadad2e81bda22e11b25f192f1680d34520a8a} Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:45.570941 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-8xrbh" event=&{ID:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748 Type:ContainerStarted Data:da2062678f72ca3bca86691478453ec531e9372503c61018f96ef5bba86dbf23} Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:45.570991 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-8xrbh" event=&{ID:e04a47b4-aa8e-4d4e-9d45-aa8037dcb748 Type:ContainerStarted Data:8510bb13dab2503578d194877e82efbb3c57ad0a439fe2f087985adeeca2f21b} Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:45.573578932Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:45.573995607Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:45.575455 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerStarted Data:936f593239e32a73c99f411c5dd4f6c509a35e0bc0c15d9738ea4de3b5de496a} Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:45.585604 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-2pfzf" event=&{ID:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 Type:ContainerStarted Data:0b790541b89bed33d7866e16060250fae888afd249ff6773c7e8c5298843b39f} Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:45.585629 1977 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-2pfzf" Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:45.585645 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-2pfzf" event=&{ID:1cc2aadb-14f2-4d6e-97fc-6fdab8889a63 Type:ContainerStarted Data:102ae4f3b08fda6cde21713ce3ea0aa9d714ac1683607033e84094ae48a306c4} Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:45.586807822Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/route-override\"" Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:45.613425729Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:45.613717419Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:46:45 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:45.613739328Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_ec991d9d-3786-4a3f-962a-4dd7e96214da\"" Jan 05 09:46:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:46.583864 1977 generic.go:296] "Generic (PLEG): container finished" podID=c0e47249-a5ff-460e-958d-53577387c154 containerID="936f593239e32a73c99f411c5dd4f6c509a35e0bc0c15d9738ea4de3b5de496a" exitCode=0 Jan 05 09:46:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:46.585206 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerDied Data:936f593239e32a73c99f411c5dd4f6c509a35e0bc0c15d9738ea4de3b5de496a} Jan 05 09:46:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:46.585618437Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451" id=5d7f8333-8272-496c-8ef0-9c2d9673e4bd name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:46.586031228Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:56f4df3c126a569448a12d5112b7c20e99e219b2b2fe6a04c4ba743396fd7751,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451],Size_:445620510,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=5d7f8333-8272-496c-8ef0-9c2d9673e4bd name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:46.587179399Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451" id=38806a8b-dae1-4dd9-8a62-7d451e05ea21 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:46.587357656Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:56f4df3c126a569448a12d5112b7c20e99e219b2b2fe6a04c4ba743396fd7751,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451],Size_:445620510,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=38806a8b-dae1-4dd9-8a62-7d451e05ea21 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:46.587876285Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/whereabouts-cni-bincopy" id=25743c62-e7ee-4a44-9d17-a16263d49101 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:46.588030706Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:46:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:46.763867838Z" level=info msg="Created container 53a63665a5f032b6a8c29f4818a61fd5d85863492c68788eb1e2a6330a60535e: openshift-multus/multus-additional-cni-plugins-5fgxc/whereabouts-cni-bincopy" id=25743c62-e7ee-4a44-9d17-a16263d49101 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:46.764281848Z" level=info msg="Starting container: 53a63665a5f032b6a8c29f4818a61fd5d85863492c68788eb1e2a6330a60535e" id=3ef3f579-7acd-4701-9f55-057d880674b7 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:46:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:46.771096173Z" level=info msg="Started container" PID=4272 containerID=53a63665a5f032b6a8c29f4818a61fd5d85863492c68788eb1e2a6330a60535e description=openshift-multus/multus-additional-cni-plugins-5fgxc/whereabouts-cni-bincopy id=3ef3f579-7acd-4701-9f55-057d880674b7 name=/runtime.v1.RuntimeService/StartContainer sandboxID=0acf147f6c1cead086fc9a6bc2a1864cb87fecc6e2901c79505a3a3c2d4136d8 Jan 05 09:46:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:46.780674799Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/upgrade_ec911136-68e2-4502-9592-008be5f7ad68\"" Jan 05 09:46:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:46.790797421Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:46:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:46.790819302Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:46:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:46.870620274Z" level=info msg="CNI monitoring event CREATE \"/var/lib/cni/bin/whereabouts\"" Jan 05 09:46:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:46.884665165Z" level=info msg="Found CNI network multus-cni-network (type=multus) at /etc/kubernetes/cni/net.d/00-multus.conf" Jan 05 09:46:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:46.884698797Z" level=info msg="Updated default CNI network name to multus-cni-network" Jan 05 09:46:46 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:46.884721725Z" level=info msg="CNI monitoring event REMOVE \"/var/lib/cni/bin/upgrade_ec911136-68e2-4502-9592-008be5f7ad68\"" Jan 05 09:46:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:47.370488 1977 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeReady" Jan 05 09:46:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:47.587655 1977 generic.go:296] "Generic (PLEG): container finished" podID=c0e47249-a5ff-460e-958d-53577387c154 containerID="53a63665a5f032b6a8c29f4818a61fd5d85863492c68788eb1e2a6330a60535e" exitCode=0 Jan 05 09:46:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:47.587740 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerDied Data:53a63665a5f032b6a8c29f4818a61fd5d85863492c68788eb1e2a6330a60535e} Jan 05 09:46:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:47.588536849Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451" id=17bc3ed4-c44f-447d-8ca6-1368690db4e2 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:47.588768144Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:56f4df3c126a569448a12d5112b7c20e99e219b2b2fe6a04c4ba743396fd7751,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451],Size_:445620510,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=17bc3ed4-c44f-447d-8ca6-1368690db4e2 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:47.589706526Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451" id=1befdd53-8084-4dd4-b388-f1d988960566 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:47.589867799Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:56f4df3c126a569448a12d5112b7c20e99e219b2b2fe6a04c4ba743396fd7751,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:161af1ae53727393d319afc77a9aadaf4faaf0c6856c0f1c2c3a01c2bd226451],Size_:445620510,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=1befdd53-8084-4dd4-b388-f1d988960566 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:47.590677871Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/whereabouts-cni" id=48d4c017-4cc7-42de-9765-14f0dd062509 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:47.590784033Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:46:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:47.764402876Z" level=info msg="Created container 48b289067ff17367d9d2a889648f112d5afe0ead56ac1d6961f1f4594a62771e: openshift-multus/multus-additional-cni-plugins-5fgxc/whereabouts-cni" id=48d4c017-4cc7-42de-9765-14f0dd062509 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:47.764923787Z" level=info msg="Starting container: 48b289067ff17367d9d2a889648f112d5afe0ead56ac1d6961f1f4594a62771e" id=874dc51f-37a2-4c43-8a48-ae535d1b3e3c name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:46:47 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:47.775411254Z" level=info msg="Started container" PID=4342 containerID=48b289067ff17367d9d2a889648f112d5afe0ead56ac1d6961f1f4594a62771e description=openshift-multus/multus-additional-cni-plugins-5fgxc/whereabouts-cni id=874dc51f-37a2-4c43-8a48-ae535d1b3e3c name=/runtime.v1.RuntimeService/StartContainer sandboxID=0acf147f6c1cead086fc9a6bc2a1864cb87fecc6e2901c79505a3a3c2d4136d8 Jan 05 09:46:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:48.591645 1977 generic.go:296] "Generic (PLEG): container finished" podID=c0e47249-a5ff-460e-958d-53577387c154 containerID="48b289067ff17367d9d2a889648f112d5afe0ead56ac1d6961f1f4594a62771e" exitCode=0 Jan 05 09:46:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:48.591682 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerDied Data:48b289067ff17367d9d2a889648f112d5afe0ead56ac1d6961f1f4594a62771e} Jan 05 09:46:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:48.592336016Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=a333753f-7a05-4ba2-81af-82ad5144873c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:48.592568292Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=a333753f-7a05-4ba2-81af-82ad5144873c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:48.593315297Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=eb0d96ad-51e4-41d6-89e6-9051fad9aa55 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:48.593477079Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=eb0d96ad-51e4-41d6-89e6-9051fad9aa55 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:46:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:48.594051592Z" level=info msg="Creating container: openshift-multus/multus-additional-cni-plugins-5fgxc/kube-multus-additional-cni-plugins" id=7ad7cea2-8f22-4434-9f7b-425ac92b97a2 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:48.594164514Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:46:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:48.754169347Z" level=info msg="Created container c003f6768d0ce88081ed645ea51f08f7e2613aa84aa9a1357785118033b0eb93: openshift-multus/multus-additional-cni-plugins-5fgxc/kube-multus-additional-cni-plugins" id=7ad7cea2-8f22-4434-9f7b-425ac92b97a2 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:46:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:48.754580852Z" level=info msg="Starting container: c003f6768d0ce88081ed645ea51f08f7e2613aa84aa9a1357785118033b0eb93" id=c66d2e4e-7cff-4f21-a5de-a12ef91afe16 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:46:48 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:46:48.761480295Z" level=info msg="Started container" PID=4405 containerID=c003f6768d0ce88081ed645ea51f08f7e2613aa84aa9a1357785118033b0eb93 description=openshift-multus/multus-additional-cni-plugins-5fgxc/kube-multus-additional-cni-plugins id=c66d2e4e-7cff-4f21-a5de-a12ef91afe16 name=/runtime.v1.RuntimeService/StartContainer sandboxID=0acf147f6c1cead086fc9a6bc2a1864cb87fecc6e2901c79505a3a3c2d4136d8 Jan 05 09:46:49 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:49.595676 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5fgxc" event=&{ID:c0e47249-a5ff-460e-958d-53577387c154 Type:ContainerStarted Data:c003f6768d0ce88081ed645ea51f08f7e2613aa84aa9a1357785118033b0eb93} Jan 05 09:46:56 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:56.390538 1977 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-2pfzf" Jan 05 09:46:57 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:46:57.698395 1977 kubelet_node_status.go:590] "Recording event message for node" node="pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl" event="NodeSchedulable" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.561771 1977 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-pqwkf] Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.561821 1977 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.588626 1977 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc] Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.588666 1977 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.603765 1977 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-network-diagnostics/network-check-source-6775957967-7tm54] Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.603800 1977 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.673980 1977 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2] Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.674011 1977 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.685436 1977 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-ingress/router-default-5cf6bbdff9-4vdhd] Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.685471 1977 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.731637 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7drzc\" (UniqueName: \"kubernetes.io/projected/85256957-515d-46bc-801a-4024088469ed-kube-api-access-7drzc\") pod \"network-check-source-6775957967-7tm54\" (UID: \"85256957-515d-46bc-801a-4024088469ed\") " pod="openshift-network-diagnostics/network-check-source-6775957967-7tm54" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.731682 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/3b3a8ee0-19b5-43a0-b793-cb6e390cba80-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-6f4f5fb7cf-xg9rc\" (UID: \"3b3a8ee0-19b5-43a0-b793-cb6e390cba80\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.731715 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/3b3a8ee0-19b5-43a0-b793-cb6e390cba80-secret-grpc-tls\") pod \"thanos-querier-6f4f5fb7cf-xg9rc\" (UID: \"3b3a8ee0-19b5-43a0-b793-cb6e390cba80\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.731791 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/3b3a8ee0-19b5-43a0-b793-cb6e390cba80-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-6f4f5fb7cf-xg9rc\" (UID: \"3b3a8ee0-19b5-43a0-b793-cb6e390cba80\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.731829 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-oauth-cookie\" (UniqueName: \"kubernetes.io/secret/3b3a8ee0-19b5-43a0-b793-cb6e390cba80-secret-thanos-querier-oauth-cookie\") pod \"thanos-querier-6f4f5fb7cf-xg9rc\" (UID: \"3b3a8ee0-19b5-43a0-b793-cb6e390cba80\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.731880 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/3b3a8ee0-19b5-43a0-b793-cb6e390cba80-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-6f4f5fb7cf-xg9rc\" (UID: \"3b3a8ee0-19b5-43a0-b793-cb6e390cba80\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.731919 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/3b3a8ee0-19b5-43a0-b793-cb6e390cba80-metrics-client-ca\") pod \"thanos-querier-6f4f5fb7cf-xg9rc\" (UID: \"3b3a8ee0-19b5-43a0-b793-cb6e390cba80\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.731978 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/f7f51e23-4c42-4696-b073-e60655077a47-tls-certificates\") pod \"prometheus-operator-admission-webhook-598468bb49-pqwkf\" (UID: \"f7f51e23-4c42-4696-b073-e60655077a47\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-pqwkf" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.732016 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/3b3a8ee0-19b5-43a0-b793-cb6e390cba80-secret-thanos-querier-tls\") pod \"thanos-querier-6f4f5fb7cf-xg9rc\" (UID: \"3b3a8ee0-19b5-43a0-b793-cb6e390cba80\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.732062 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-querier-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3b3a8ee0-19b5-43a0-b793-cb6e390cba80-thanos-querier-trusted-ca-bundle\") pod \"thanos-querier-6f4f5fb7cf-xg9rc\" (UID: \"3b3a8ee0-19b5-43a0-b793-cb6e390cba80\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.732095 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfwrf\" (UniqueName: \"kubernetes.io/projected/3b3a8ee0-19b5-43a0-b793-cb6e390cba80-kube-api-access-qfwrf\") pod \"thanos-querier-6f4f5fb7cf-xg9rc\" (UID: \"3b3a8ee0-19b5-43a0-b793-cb6e390cba80\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.832575 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6b28bc9-5526-449d-b81c-ba4c654e254b-config\") pod \"prometheus-adapter-5894cc86c7-rxkb2\" (UID: \"a6b28bc9-5526-449d-b81c-ba4c654e254b\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.832627 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/f7f51e23-4c42-4696-b073-e60655077a47-tls-certificates\") pod \"prometheus-operator-admission-webhook-598468bb49-pqwkf\" (UID: \"f7f51e23-4c42-4696-b073-e60655077a47\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-pqwkf" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.832659 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/3b3a8ee0-19b5-43a0-b793-cb6e390cba80-metrics-client-ca\") pod \"thanos-querier-6f4f5fb7cf-xg9rc\" (UID: \"3b3a8ee0-19b5-43a0-b793-cb6e390cba80\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.832696 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-adapter-prometheus-config\" (UniqueName: \"kubernetes.io/configmap/a6b28bc9-5526-449d-b81c-ba4c654e254b-prometheus-adapter-prometheus-config\") pod \"prometheus-adapter-5894cc86c7-rxkb2\" (UID: \"a6b28bc9-5526-449d-b81c-ba4c654e254b\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.832727 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/78fead4b-9b2f-4ed6-a4f7-4c8479e076ee-stats-auth\") pod \"router-default-5cf6bbdff9-4vdhd\" (UID: \"78fead4b-9b2f-4ed6-a4f7-4c8479e076ee\") " pod="openshift-ingress/router-default-5cf6bbdff9-4vdhd" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.832753 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls\" (UniqueName: \"kubernetes.io/secret/a6b28bc9-5526-449d-b81c-ba4c654e254b-tls\") pod \"prometheus-adapter-5894cc86c7-rxkb2\" (UID: \"a6b28bc9-5526-449d-b81c-ba4c654e254b\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.832789 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/3b3a8ee0-19b5-43a0-b793-cb6e390cba80-secret-thanos-querier-tls\") pod \"thanos-querier-6f4f5fb7cf-xg9rc\" (UID: \"3b3a8ee0-19b5-43a0-b793-cb6e390cba80\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.832829 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a6b28bc9-5526-449d-b81c-ba4c654e254b-serving-certs-ca-bundle\") pod \"prometheus-adapter-5894cc86c7-rxkb2\" (UID: \"a6b28bc9-5526-449d-b81c-ba4c654e254b\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.832881 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjm4j\" (UniqueName: \"kubernetes.io/projected/78fead4b-9b2f-4ed6-a4f7-4c8479e076ee-kube-api-access-bjm4j\") pod \"router-default-5cf6bbdff9-4vdhd\" (UID: \"78fead4b-9b2f-4ed6-a4f7-4c8479e076ee\") " pod="openshift-ingress/router-default-5cf6bbdff9-4vdhd" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.832917 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/a6b28bc9-5526-449d-b81c-ba4c654e254b-audit-log\") pod \"prometheus-adapter-5894cc86c7-rxkb2\" (UID: \"a6b28bc9-5526-449d-b81c-ba4c654e254b\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.832970 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-adapter-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/a6b28bc9-5526-449d-b81c-ba4c654e254b-prometheus-adapter-audit-profiles\") pod \"prometheus-adapter-5894cc86c7-rxkb2\" (UID: \"a6b28bc9-5526-449d-b81c-ba4c654e254b\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.833008 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-qfwrf\" (UniqueName: \"kubernetes.io/projected/3b3a8ee0-19b5-43a0-b793-cb6e390cba80-kube-api-access-qfwrf\") pod \"thanos-querier-6f4f5fb7cf-xg9rc\" (UID: \"3b3a8ee0-19b5-43a0-b793-cb6e390cba80\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.833050 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"thanos-querier-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3b3a8ee0-19b5-43a0-b793-cb6e390cba80-thanos-querier-trusted-ca-bundle\") pod \"thanos-querier-6f4f5fb7cf-xg9rc\" (UID: \"3b3a8ee0-19b5-43a0-b793-cb6e390cba80\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.833088 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvsvt\" (UniqueName: \"kubernetes.io/projected/a6b28bc9-5526-449d-b81c-ba4c654e254b-kube-api-access-jvsvt\") pod \"prometheus-adapter-5894cc86c7-rxkb2\" (UID: \"a6b28bc9-5526-449d-b81c-ba4c654e254b\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.833124 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-7drzc\" (UniqueName: \"kubernetes.io/projected/85256957-515d-46bc-801a-4024088469ed-kube-api-access-7drzc\") pod \"network-check-source-6775957967-7tm54\" (UID: \"85256957-515d-46bc-801a-4024088469ed\") " pod="openshift-network-diagnostics/network-check-source-6775957967-7tm54" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.833156 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/78fead4b-9b2f-4ed6-a4f7-4c8479e076ee-default-certificate\") pod \"router-default-5cf6bbdff9-4vdhd\" (UID: \"78fead4b-9b2f-4ed6-a4f7-4c8479e076ee\") " pod="openshift-ingress/router-default-5cf6bbdff9-4vdhd" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.833190 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/78fead4b-9b2f-4ed6-a4f7-4c8479e076ee-metrics-certs\") pod \"router-default-5cf6bbdff9-4vdhd\" (UID: \"78fead4b-9b2f-4ed6-a4f7-4c8479e076ee\") " pod="openshift-ingress/router-default-5cf6bbdff9-4vdhd" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.833227 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/3b3a8ee0-19b5-43a0-b793-cb6e390cba80-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-6f4f5fb7cf-xg9rc\" (UID: \"3b3a8ee0-19b5-43a0-b793-cb6e390cba80\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.833264 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/3b3a8ee0-19b5-43a0-b793-cb6e390cba80-secret-grpc-tls\") pod \"thanos-querier-6f4f5fb7cf-xg9rc\" (UID: \"3b3a8ee0-19b5-43a0-b793-cb6e390cba80\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.833301 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/3b3a8ee0-19b5-43a0-b793-cb6e390cba80-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-6f4f5fb7cf-xg9rc\" (UID: \"3b3a8ee0-19b5-43a0-b793-cb6e390cba80\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.833334 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a6b28bc9-5526-449d-b81c-ba4c654e254b-tmpfs\") pod \"prometheus-adapter-5894cc86c7-rxkb2\" (UID: \"a6b28bc9-5526-449d-b81c-ba4c654e254b\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.833370 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-oauth-cookie\" (UniqueName: \"kubernetes.io/secret/3b3a8ee0-19b5-43a0-b793-cb6e390cba80-secret-thanos-querier-oauth-cookie\") pod \"thanos-querier-6f4f5fb7cf-xg9rc\" (UID: \"3b3a8ee0-19b5-43a0-b793-cb6e390cba80\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.833407 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/3b3a8ee0-19b5-43a0-b793-cb6e390cba80-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-6f4f5fb7cf-xg9rc\" (UID: \"3b3a8ee0-19b5-43a0-b793-cb6e390cba80\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.833440 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/78fead4b-9b2f-4ed6-a4f7-4c8479e076ee-service-ca-bundle\") pod \"router-default-5cf6bbdff9-4vdhd\" (UID: \"78fead4b-9b2f-4ed6-a4f7-4c8479e076ee\") " pod="openshift-ingress/router-default-5cf6bbdff9-4vdhd" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.834423 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"thanos-querier-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3b3a8ee0-19b5-43a0-b793-cb6e390cba80-thanos-querier-trusted-ca-bundle\") pod \"thanos-querier-6f4f5fb7cf-xg9rc\" (UID: \"3b3a8ee0-19b5-43a0-b793-cb6e390cba80\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.835049 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/3b3a8ee0-19b5-43a0-b793-cb6e390cba80-metrics-client-ca\") pod \"thanos-querier-6f4f5fb7cf-xg9rc\" (UID: \"3b3a8ee0-19b5-43a0-b793-cb6e390cba80\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.837362 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/f7f51e23-4c42-4696-b073-e60655077a47-tls-certificates\") pod \"prometheus-operator-admission-webhook-598468bb49-pqwkf\" (UID: \"f7f51e23-4c42-4696-b073-e60655077a47\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-pqwkf" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.841317 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/3b3a8ee0-19b5-43a0-b793-cb6e390cba80-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-6f4f5fb7cf-xg9rc\" (UID: \"3b3a8ee0-19b5-43a0-b793-cb6e390cba80\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.843447 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/3b3a8ee0-19b5-43a0-b793-cb6e390cba80-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-6f4f5fb7cf-xg9rc\" (UID: \"3b3a8ee0-19b5-43a0-b793-cb6e390cba80\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.845875 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/3b3a8ee0-19b5-43a0-b793-cb6e390cba80-secret-grpc-tls\") pod \"thanos-querier-6f4f5fb7cf-xg9rc\" (UID: \"3b3a8ee0-19b5-43a0-b793-cb6e390cba80\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.848046 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/3b3a8ee0-19b5-43a0-b793-cb6e390cba80-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-6f4f5fb7cf-xg9rc\" (UID: \"3b3a8ee0-19b5-43a0-b793-cb6e390cba80\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.851027 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/3b3a8ee0-19b5-43a0-b793-cb6e390cba80-secret-thanos-querier-tls\") pod \"thanos-querier-6f4f5fb7cf-xg9rc\" (UID: \"3b3a8ee0-19b5-43a0-b793-cb6e390cba80\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.851377 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-oauth-cookie\" (UniqueName: \"kubernetes.io/secret/3b3a8ee0-19b5-43a0-b793-cb6e390cba80-secret-thanos-querier-oauth-cookie\") pod \"thanos-querier-6f4f5fb7cf-xg9rc\" (UID: \"3b3a8ee0-19b5-43a0-b793-cb6e390cba80\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.874459 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfwrf\" (UniqueName: \"kubernetes.io/projected/3b3a8ee0-19b5-43a0-b793-cb6e390cba80-kube-api-access-qfwrf\") pod \"thanos-querier-6f4f5fb7cf-xg9rc\" (UID: \"3b3a8ee0-19b5-43a0-b793-cb6e390cba80\") " pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.877111 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-7drzc\" (UniqueName: \"kubernetes.io/projected/85256957-515d-46bc-801a-4024088469ed-kube-api-access-7drzc\") pod \"network-check-source-6775957967-7tm54\" (UID: \"85256957-515d-46bc-801a-4024088469ed\") " pod="openshift-network-diagnostics/network-check-source-6775957967-7tm54" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.904086 1977 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-pqwkf" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:04.904858850Z" level=info msg="Running pod sandbox: openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-pqwkf/POD" id=c08d6917-dd19-44ad-9a33-db38f0bec039 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:04.904921675Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:04.919620449Z" level=info msg="Got pod network &{Name:prometheus-operator-admission-webhook-598468bb49-pqwkf Namespace:openshift-monitoring ID:31719a55b8c87ba34a4133b598d3eb7b108f583c8f149bd000793f3743c53086 UID:f7f51e23-4c42-4696-b073-e60655077a47 NetNS:/var/run/netns/b7c7b6da-a998-4425-a6b6-867f29908b8c Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:04.919648281Z" level=info msg="Adding pod openshift-monitoring_prometheus-operator-admission-webhook-598468bb49-pqwkf to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.934188 1977 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.934265 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"prometheus-adapter-prometheus-config\" (UniqueName: \"kubernetes.io/configmap/a6b28bc9-5526-449d-b81c-ba4c654e254b-prometheus-adapter-prometheus-config\") pod \"prometheus-adapter-5894cc86c7-rxkb2\" (UID: \"a6b28bc9-5526-449d-b81c-ba4c654e254b\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.934304 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/78fead4b-9b2f-4ed6-a4f7-4c8479e076ee-stats-auth\") pod \"router-default-5cf6bbdff9-4vdhd\" (UID: \"78fead4b-9b2f-4ed6-a4f7-4c8479e076ee\") " pod="openshift-ingress/router-default-5cf6bbdff9-4vdhd" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.934332 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tls\" (UniqueName: \"kubernetes.io/secret/a6b28bc9-5526-449d-b81c-ba4c654e254b-tls\") pod \"prometheus-adapter-5894cc86c7-rxkb2\" (UID: \"a6b28bc9-5526-449d-b81c-ba4c654e254b\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.934368 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a6b28bc9-5526-449d-b81c-ba4c654e254b-serving-certs-ca-bundle\") pod \"prometheus-adapter-5894cc86c7-rxkb2\" (UID: \"a6b28bc9-5526-449d-b81c-ba4c654e254b\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.934401 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-bjm4j\" (UniqueName: \"kubernetes.io/projected/78fead4b-9b2f-4ed6-a4f7-4c8479e076ee-kube-api-access-bjm4j\") pod \"router-default-5cf6bbdff9-4vdhd\" (UID: \"78fead4b-9b2f-4ed6-a4f7-4c8479e076ee\") " pod="openshift-ingress/router-default-5cf6bbdff9-4vdhd" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.934421 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/a6b28bc9-5526-449d-b81c-ba4c654e254b-audit-log\") pod \"prometheus-adapter-5894cc86c7-rxkb2\" (UID: \"a6b28bc9-5526-449d-b81c-ba4c654e254b\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.934443 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"prometheus-adapter-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/a6b28bc9-5526-449d-b81c-ba4c654e254b-prometheus-adapter-audit-profiles\") pod \"prometheus-adapter-5894cc86c7-rxkb2\" (UID: \"a6b28bc9-5526-449d-b81c-ba4c654e254b\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.934469 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-jvsvt\" (UniqueName: \"kubernetes.io/projected/a6b28bc9-5526-449d-b81c-ba4c654e254b-kube-api-access-jvsvt\") pod \"prometheus-adapter-5894cc86c7-rxkb2\" (UID: \"a6b28bc9-5526-449d-b81c-ba4c654e254b\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.934489 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/78fead4b-9b2f-4ed6-a4f7-4c8479e076ee-default-certificate\") pod \"router-default-5cf6bbdff9-4vdhd\" (UID: \"78fead4b-9b2f-4ed6-a4f7-4c8479e076ee\") " pod="openshift-ingress/router-default-5cf6bbdff9-4vdhd" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.934507 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/78fead4b-9b2f-4ed6-a4f7-4c8479e076ee-metrics-certs\") pod \"router-default-5cf6bbdff9-4vdhd\" (UID: \"78fead4b-9b2f-4ed6-a4f7-4c8479e076ee\") " pod="openshift-ingress/router-default-5cf6bbdff9-4vdhd" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.934532 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a6b28bc9-5526-449d-b81c-ba4c654e254b-tmpfs\") pod \"prometheus-adapter-5894cc86c7-rxkb2\" (UID: \"a6b28bc9-5526-449d-b81c-ba4c654e254b\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.934554 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/78fead4b-9b2f-4ed6-a4f7-4c8479e076ee-service-ca-bundle\") pod \"router-default-5cf6bbdff9-4vdhd\" (UID: \"78fead4b-9b2f-4ed6-a4f7-4c8479e076ee\") " pod="openshift-ingress/router-default-5cf6bbdff9-4vdhd" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.934575 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6b28bc9-5526-449d-b81c-ba4c654e254b-config\") pod \"prometheus-adapter-5894cc86c7-rxkb2\" (UID: \"a6b28bc9-5526-449d-b81c-ba4c654e254b\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.935076 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6b28bc9-5526-449d-b81c-ba4c654e254b-config\") pod \"prometheus-adapter-5894cc86c7-rxkb2\" (UID: \"a6b28bc9-5526-449d-b81c-ba4c654e254b\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.935420 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"prometheus-adapter-prometheus-config\" (UniqueName: \"kubernetes.io/configmap/a6b28bc9-5526-449d-b81c-ba4c654e254b-prometheus-adapter-prometheus-config\") pod \"prometheus-adapter-5894cc86c7-rxkb2\" (UID: \"a6b28bc9-5526-449d-b81c-ba4c654e254b\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.938219 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"prometheus-adapter-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/a6b28bc9-5526-449d-b81c-ba4c654e254b-prometheus-adapter-audit-profiles\") pod \"prometheus-adapter-5894cc86c7-rxkb2\" (UID: \"a6b28bc9-5526-449d-b81c-ba4c654e254b\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.938811 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a6b28bc9-5526-449d-b81c-ba4c654e254b-serving-certs-ca-bundle\") pod \"prometheus-adapter-5894cc86c7-rxkb2\" (UID: \"a6b28bc9-5526-449d-b81c-ba4c654e254b\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.939305 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/a6b28bc9-5526-449d-b81c-ba4c654e254b-audit-log\") pod \"prometheus-adapter-5894cc86c7-rxkb2\" (UID: \"a6b28bc9-5526-449d-b81c-ba4c654e254b\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.940018 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a6b28bc9-5526-449d-b81c-ba4c654e254b-tmpfs\") pod \"prometheus-adapter-5894cc86c7-rxkb2\" (UID: \"a6b28bc9-5526-449d-b81c-ba4c654e254b\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.940500 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/78fead4b-9b2f-4ed6-a4f7-4c8479e076ee-service-ca-bundle\") pod \"router-default-5cf6bbdff9-4vdhd\" (UID: \"78fead4b-9b2f-4ed6-a4f7-4c8479e076ee\") " pod="openshift-ingress/router-default-5cf6bbdff9-4vdhd" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.941175 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/78fead4b-9b2f-4ed6-a4f7-4c8479e076ee-stats-auth\") pod \"router-default-5cf6bbdff9-4vdhd\" (UID: \"78fead4b-9b2f-4ed6-a4f7-4c8479e076ee\") " pod="openshift-ingress/router-default-5cf6bbdff9-4vdhd" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.943648 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tls\" (UniqueName: \"kubernetes.io/secret/a6b28bc9-5526-449d-b81c-ba4c654e254b-tls\") pod \"prometheus-adapter-5894cc86c7-rxkb2\" (UID: \"a6b28bc9-5526-449d-b81c-ba4c654e254b\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.945743 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/78fead4b-9b2f-4ed6-a4f7-4c8479e076ee-metrics-certs\") pod \"router-default-5cf6bbdff9-4vdhd\" (UID: \"78fead4b-9b2f-4ed6-a4f7-4c8479e076ee\") " pod="openshift-ingress/router-default-5cf6bbdff9-4vdhd" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:04.946418721Z" level=info msg="Running pod sandbox: openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc/POD" id=08a724bc-6159-4460-a395-e7d366f39a93 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:04.946482216Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.948545 1977 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-6775957967-7tm54" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:04.950019555Z" level=info msg="Running pod sandbox: openshift-network-diagnostics/network-check-source-6775957967-7tm54/POD" id=05035b9b-126d-4087-a152-381e6a2ddd10 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:04.950068112Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.950130 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/78fead4b-9b2f-4ed6-a4f7-4c8479e076ee-default-certificate\") pod \"router-default-5cf6bbdff9-4vdhd\" (UID: \"78fead4b-9b2f-4ed6-a4f7-4c8479e076ee\") " pod="openshift-ingress/router-default-5cf6bbdff9-4vdhd" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.962097 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvsvt\" (UniqueName: \"kubernetes.io/projected/a6b28bc9-5526-449d-b81c-ba4c654e254b-kube-api-access-jvsvt\") pod \"prometheus-adapter-5894cc86c7-rxkb2\" (UID: \"a6b28bc9-5526-449d-b81c-ba4c654e254b\") " pod="openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2" Jan 05 09:47:04 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:04.963932 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjm4j\" (UniqueName: \"kubernetes.io/projected/78fead4b-9b2f-4ed6-a4f7-4c8479e076ee-kube-api-access-bjm4j\") pod \"router-default-5cf6bbdff9-4vdhd\" (UID: \"78fead4b-9b2f-4ed6-a4f7-4c8479e076ee\") " pod="openshift-ingress/router-default-5cf6bbdff9-4vdhd" Jan 05 09:47:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:05.003595 1977 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2" Jan 05 09:47:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:05.004281315Z" level=info msg="Running pod sandbox: openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2/POD" id=3136748b-a44e-4844-8f78-e9ead12d9921 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:47:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:05.004335424Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:47:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:05.013813 1977 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5cf6bbdff9-4vdhd" Jan 05 09:47:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:05.019069553Z" level=info msg="Running pod sandbox: openshift-ingress/router-default-5cf6bbdff9-4vdhd/POD" id=e4df48a1-da4f-46af-9abd-0758b8d66857 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:47:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:05.019133420Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:47:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:05.040066924Z" level=info msg="Got pod network &{Name:thanos-querier-6f4f5fb7cf-xg9rc Namespace:openshift-monitoring ID:54dbaba32104ac88a1e3ed7c1c5c8cd5be28a9c930620bf99f4e31003ced75b4 UID:3b3a8ee0-19b5-43a0-b793-cb6e390cba80 NetNS:/var/run/netns/3f224775-9d68-4683-9226-aea6453b8723 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:47:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:05.040109595Z" level=info msg="Adding pod openshift-monitoring_thanos-querier-6f4f5fb7cf-xg9rc to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:47:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:05.050198959Z" level=info msg="Got pod network &{Name:network-check-source-6775957967-7tm54 Namespace:openshift-network-diagnostics ID:dd5d6f1a8a242dced9e946b55763f6bdd8ee37d23cb679a2b3b0dde79120ed24 UID:85256957-515d-46bc-801a-4024088469ed NetNS:/var/run/netns/9d427379-59b3-4668-8688-8709667cb006 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:47:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:05.050246482Z" level=info msg="Adding pod openshift-network-diagnostics_network-check-source-6775957967-7tm54 to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:47:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:05.140978410Z" level=info msg="Got pod network &{Name:prometheus-adapter-5894cc86c7-rxkb2 Namespace:openshift-monitoring ID:ad84fb5151c7b562c151f9b22fc3980b8cb0e4f105022cec531ada830a445264 UID:a6b28bc9-5526-449d-b81c-ba4c654e254b NetNS:/var/run/netns/af514492-5b7e-435d-be41-713c63ea07ae Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:47:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:05.141016430Z" level=info msg="Adding pod openshift-monitoring_prometheus-adapter-5894cc86c7-rxkb2 to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:47:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:05.149130871Z" level=info msg="Got pod network &{Name:router-default-5cf6bbdff9-4vdhd Namespace:openshift-ingress ID:4660d5572cae6cef99bbb174a92ded51ecce6bfa33a8aede8e8d313ee080384d UID:78fead4b-9b2f-4ed6-a4f7-4c8479e076ee NetNS:/var/run/netns/dbfdeaf8-528d-4e2e-b7d2-9087124a29f7 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:47:05 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:05.149163465Z" level=info msg="Adding pod openshift-ingress_router-default-5cf6bbdff9-4vdhd to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:06.262787 1977 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc] Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: 2023-01-05T09:47:06Z [verbose] Add: openshift-monitoring:thanos-querier-6f4f5fb7cf-xg9rc:3b3a8ee0-19b5-43a0-b793-cb6e390cba80:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/3f224775-9d68-4683-9226-aea6453b8723"}],"ips":[{"version":"4","interface":0,"address":"10.131.0.7/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: I0105 09:47:06.235544 4718 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"thanos-querier-6f4f5fb7cf-xg9rc", UID:"3b3a8ee0-19b5-43a0-b793-cb6e390cba80", APIVersion:"v1", ResourceVersion:"107738", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.7/23] from openshift-sdn Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.263938604Z" level=info msg="Got pod network &{Name:thanos-querier-6f4f5fb7cf-xg9rc Namespace:openshift-monitoring ID:54dbaba32104ac88a1e3ed7c1c5c8cd5be28a9c930620bf99f4e31003ced75b4 UID:3b3a8ee0-19b5-43a0-b793-cb6e390cba80 NetNS:/var/run/netns/3f224775-9d68-4683-9226-aea6453b8723 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.264119883Z" level=info msg="Checking pod openshift-monitoring_thanos-querier-6f4f5fb7cf-xg9rc for CNI network multus-cni-network (type=multus)" Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: W0105 09:47:06.266375 1977 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b3a8ee0_19b5_43a0_b793_cb6e390cba80.slice/crio-54dbaba32104ac88a1e3ed7c1c5c8cd5be28a9c930620bf99f4e31003ced75b4.scope WatchSource:0}: Error finding container 54dbaba32104ac88a1e3ed7c1c5c8cd5be28a9c930620bf99f4e31003ced75b4: Status 404 returned error can't find the container with id 54dbaba32104ac88a1e3ed7c1c5c8cd5be28a9c930620bf99f4e31003ced75b4 Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.278218664Z" level=info msg="Ran pod sandbox 54dbaba32104ac88a1e3ed7c1c5c8cd5be28a9c930620bf99f4e31003ced75b4 with infra container: openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc/POD" id=08a724bc-6159-4460-a395-e7d366f39a93 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.280546376Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc" id=6a352e0d-f8ae-41ec-a71a-99a83d3dda2b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.280727446Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:08ff7a43d9185529520a29e114f0c28b070000b464a44b347240015f3b466488,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc],Size_:368278089,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=6a352e0d-f8ae-41ec-a71a-99a83d3dda2b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.281385981Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc" id=f2a4984c-6df6-4ae5-8a9a-486730955855 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.281581399Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:08ff7a43d9185529520a29e114f0c28b070000b464a44b347240015f3b466488,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc],Size_:368278089,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=f2a4984c-6df6-4ae5-8a9a-486730955855 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.282769888Z" level=info msg="Creating container: openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc/thanos-query" id=7264399a-c3e2-472c-8eb5-c603062f5bd7 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.282901599Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.485675631Z" level=info msg="Created container 1b58503d61b261b496c618ff3caf6c8bc7cb377fceeb8da0481baf87116c6831: openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc/thanos-query" id=7264399a-c3e2-472c-8eb5-c603062f5bd7 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.486563258Z" level=info msg="Starting container: 1b58503d61b261b496c618ff3caf6c8bc7cb377fceeb8da0481baf87116c6831" id=74a656ab-fd98-4d67-8aa7-2985051add4c name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.508681319Z" level=info msg="Started container" PID=4886 containerID=1b58503d61b261b496c618ff3caf6c8bc7cb377fceeb8da0481baf87116c6831 description=openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc/thanos-query id=74a656ab-fd98-4d67-8aa7-2985051add4c name=/runtime.v1.RuntimeService/StartContainer sandboxID=54dbaba32104ac88a1e3ed7c1c5c8cd5be28a9c930620bf99f4e31003ced75b4 Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.528695682Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=0446d608-5d95-4cfd-9585-147e44426631 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.528991111Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=0446d608-5d95-4cfd-9585-147e44426631 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.532937882Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=16387f63-24bd-40d7-b387-35873918197e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.533194389Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=16387f63-24bd-40d7-b387-35873918197e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.535048978Z" level=info msg="Creating container: openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc/oauth-proxy" id=4c98c78b-e142-47ec-afda-a6b84f1cb58f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.535177264Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:06.620712 1977 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2] Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: 2023-01-05T09:47:06Z [verbose] Add: openshift-monitoring:prometheus-adapter-5894cc86c7-rxkb2:a6b28bc9-5526-449d-b81c-ba4c654e254b:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/af514492-5b7e-435d-be41-713c63ea07ae"}],"ips":[{"version":"4","interface":0,"address":"10.131.0.9/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: I0105 09:47:06.593603 4755 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"prometheus-adapter-5894cc86c7-rxkb2", UID:"a6b28bc9-5526-449d-b81c-ba4c654e254b", APIVersion:"v1", ResourceVersion:"107762", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.9/23] from openshift-sdn Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.633448758Z" level=info msg="Got pod network &{Name:prometheus-adapter-5894cc86c7-rxkb2 Namespace:openshift-monitoring ID:ad84fb5151c7b562c151f9b22fc3980b8cb0e4f105022cec531ada830a445264 UID:a6b28bc9-5526-449d-b81c-ba4c654e254b NetNS:/var/run/netns/af514492-5b7e-435d-be41-713c63ea07ae Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.633650744Z" level=info msg="Checking pod openshift-monitoring_prometheus-adapter-5894cc86c7-rxkb2 for CNI network multus-cni-network (type=multus)" Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: W0105 09:47:06.638432 1977 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda6b28bc9_5526_449d_b81c_ba4c654e254b.slice/crio-ad84fb5151c7b562c151f9b22fc3980b8cb0e4f105022cec531ada830a445264.scope WatchSource:0}: Error finding container ad84fb5151c7b562c151f9b22fc3980b8cb0e4f105022cec531ada830a445264: Status 404 returned error can't find the container with id ad84fb5151c7b562c151f9b22fc3980b8cb0e4f105022cec531ada830a445264 Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:06.640559 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" event=&{ID:3b3a8ee0-19b5-43a0-b793-cb6e390cba80 Type:ContainerStarted Data:1b58503d61b261b496c618ff3caf6c8bc7cb377fceeb8da0481baf87116c6831} Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:06.640594 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" event=&{ID:3b3a8ee0-19b5-43a0-b793-cb6e390cba80 Type:ContainerStarted Data:54dbaba32104ac88a1e3ed7c1c5c8cd5be28a9c930620bf99f4e31003ced75b4} Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.650464269Z" level=info msg="Ran pod sandbox ad84fb5151c7b562c151f9b22fc3980b8cb0e4f105022cec531ada830a445264 with infra container: openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2/POD" id=3136748b-a44e-4844-8f78-e9ead12d9921 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.653217781Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9696f8f231d49aadf63ab980ec5c17c4fd87011930daecd30cc9e4ee102f1aeb" id=eefecb3f-8c4d-4e67-a061-78bcf6558146 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.653446078Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:ea0b364e3a8a4e4b3f4ab3288e42c5be8021187d32791029f9cb9a500c8da100,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9696f8f231d49aadf63ab980ec5c17c4fd87011930daecd30cc9e4ee102f1aeb],Size_:402351209,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=eefecb3f-8c4d-4e67-a061-78bcf6558146 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.662698358Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9696f8f231d49aadf63ab980ec5c17c4fd87011930daecd30cc9e4ee102f1aeb" id=b8ff990e-e769-4a01-9695-805f61b2b3ea name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.662903251Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:ea0b364e3a8a4e4b3f4ab3288e42c5be8021187d32791029f9cb9a500c8da100,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:9696f8f231d49aadf63ab980ec5c17c4fd87011930daecd30cc9e4ee102f1aeb],Size_:402351209,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b8ff990e-e769-4a01-9695-805f61b2b3ea name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.664318499Z" level=info msg="Creating container: openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2/prometheus-adapter" id=06857674-eb2c-4ea2-9e0a-b3bd38099bab name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.664436137Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.940060568Z" level=info msg="Created container d84da8bec40a36151af3705b3b83dd377e211227e824a7ad34e3a18086ec78d8: openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc/oauth-proxy" id=4c98c78b-e142-47ec-afda-a6b84f1cb58f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.942993885Z" level=info msg="Starting container: d84da8bec40a36151af3705b3b83dd377e211227e824a7ad34e3a18086ec78d8" id=50480afd-bccf-4e0e-95a4-c3fa2a0a8857 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.966433836Z" level=info msg="Started container" PID=4979 containerID=d84da8bec40a36151af3705b3b83dd377e211227e824a7ad34e3a18086ec78d8 description=openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc/oauth-proxy id=50480afd-bccf-4e0e-95a4-c3fa2a0a8857 name=/runtime.v1.RuntimeService/StartContainer sandboxID=54dbaba32104ac88a1e3ed7c1c5c8cd5be28a9c930620bf99f4e31003ced75b4 Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.995502775Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=9d2c4df7-2df8-4bba-8f5e-368b049e3afb name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.995736736Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=9d2c4df7-2df8-4bba-8f5e-368b049e3afb name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.997139091Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=95cf3b29-99b6-47bf-b92f-6208b096ba7e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:06.997330386Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=95cf3b29-99b6-47bf-b92f-6208b096ba7e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.031007394Z" level=info msg="Creating container: openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc/kube-rbac-proxy" id=b4ebb602-5c71-4bb4-9e86-f55bf115043f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.031144953Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.096164303Z" level=info msg="Created container 846ffcc9f1e0feb970619b56fe60c2f15822dda468bcd0552ca0a3625b6c58b7: openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2/prometheus-adapter" id=06857674-eb2c-4ea2-9e0a-b3bd38099bab name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.096848450Z" level=info msg="Starting container: 846ffcc9f1e0feb970619b56fe60c2f15822dda468bcd0552ca0a3625b6c58b7" id=ec6a871c-0ad7-4d36-bd2d-c6a7b0c3d314 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.115471 1977 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-pqwkf] Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: 2023-01-05T09:47:07Z [verbose] Add: openshift-monitoring:prometheus-operator-admission-webhook-598468bb49-pqwkf:f7f51e23-4c42-4696-b073-e60655077a47:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/b7c7b6da-a998-4425-a6b6-867f29908b8c"}],"ips":[{"version":"4","interface":0,"address":"10.131.0.6/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: I0105 09:47:07.079897 4686 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"prometheus-operator-admission-webhook-598468bb49-pqwkf", UID:"f7f51e23-4c42-4696-b073-e60655077a47", APIVersion:"v1", ResourceVersion:"107711", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.6/23] from openshift-sdn Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.121135524Z" level=info msg="Got pod network &{Name:prometheus-operator-admission-webhook-598468bb49-pqwkf Namespace:openshift-monitoring ID:31719a55b8c87ba34a4133b598d3eb7b108f583c8f149bd000793f3743c53086 UID:f7f51e23-4c42-4696-b073-e60655077a47 NetNS:/var/run/netns/b7c7b6da-a998-4425-a6b6-867f29908b8c Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.121246166Z" level=info msg="Checking pod openshift-monitoring_prometheus-operator-admission-webhook-598468bb49-pqwkf for CNI network multus-cni-network (type=multus)" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.132009 1977 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-network-diagnostics/network-check-source-6775957967-7tm54] Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.132773550Z" level=info msg="Started container" PID=5010 containerID=846ffcc9f1e0feb970619b56fe60c2f15822dda468bcd0552ca0a3625b6c58b7 description=openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2/prometheus-adapter id=ec6a871c-0ad7-4d36-bd2d-c6a7b0c3d314 name=/runtime.v1.RuntimeService/StartContainer sandboxID=ad84fb5151c7b562c151f9b22fc3980b8cb0e4f105022cec531ada830a445264 Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: 2023-01-05T09:47:07Z [verbose] Add: openshift-network-diagnostics:network-check-source-6775957967-7tm54:85256957-515d-46bc-801a-4024088469ed:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/9d427379-59b3-4668-8688-8709667cb006"}],"ips":[{"version":"4","interface":0,"address":"10.131.0.8/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: I0105 09:47:07.094301 4723 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-network-diagnostics", Name:"network-check-source-6775957967-7tm54", UID:"85256957-515d-46bc-801a-4024088469ed", APIVersion:"v1", ResourceVersion:"107747", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.8/23] from openshift-sdn Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.136637256Z" level=info msg="Got pod network &{Name:network-check-source-6775957967-7tm54 Namespace:openshift-network-diagnostics ID:dd5d6f1a8a242dced9e946b55763f6bdd8ee37d23cb679a2b3b0dde79120ed24 UID:85256957-515d-46bc-801a-4024088469ed NetNS:/var/run/netns/9d427379-59b3-4668-8688-8709667cb006 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.136782456Z" level=info msg="Checking pod openshift-network-diagnostics_network-check-source-6775957967-7tm54 for CNI network multus-cni-network (type=multus)" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: W0105 09:47:07.140201 1977 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf7f51e23_4c42_4696_b073_e60655077a47.slice/crio-31719a55b8c87ba34a4133b598d3eb7b108f583c8f149bd000793f3743c53086.scope WatchSource:0}: Error finding container 31719a55b8c87ba34a4133b598d3eb7b108f583c8f149bd000793f3743c53086: Status 404 returned error can't find the container with id 31719a55b8c87ba34a4133b598d3eb7b108f583c8f149bd000793f3743c53086 Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.139979276Z" level=info msg="Ran pod sandbox 31719a55b8c87ba34a4133b598d3eb7b108f583c8f149bd000793f3743c53086 with infra container: openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-pqwkf/POD" id=c08d6917-dd19-44ad-9a33-db38f0bec039 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.143104837Z" level=info msg="Ran pod sandbox dd5d6f1a8a242dced9e946b55763f6bdd8ee37d23cb679a2b3b0dde79120ed24 with infra container: openshift-network-diagnostics/network-check-source-6775957967-7tm54/POD" id=05035b9b-126d-4087-a152-381e6a2ddd10 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.147401492Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:dbf46164dcc3b2b6ef5df5d043ddc4fd859fd2c73c4862808424b3df6a1e27e7" id=dc244b56-cb4b-4050-a526-513e6a8a0fc0 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.147627803Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:20e9403b34a0d9d25151a4ddbee3f69b99894fd8e058e6cb7310fe95080ecb03,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:dbf46164dcc3b2b6ef5df5d043ddc4fd859fd2c73c4862808424b3df6a1e27e7],Size_:374119253,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=dc244b56-cb4b-4050-a526-513e6a8a0fc0 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.150012150Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7" id=afe4e73f-b22f-4139-96e0-45beaee652ec name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.150212749Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:4cb41c73a232ec4401c5db32cdb523810dc12ac605cdc6a9ce388e90930eb6c4,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7],Size_:515955107,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=afe4e73f-b22f-4139-96e0-45beaee652ec name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.150292134Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:dbf46164dcc3b2b6ef5df5d043ddc4fd859fd2c73c4862808424b3df6a1e27e7" id=0a9ef099-e20e-4fe4-96ed-6e360c5e8cec name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.150440658Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:20e9403b34a0d9d25151a4ddbee3f69b99894fd8e058e6cb7310fe95080ecb03,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:dbf46164dcc3b2b6ef5df5d043ddc4fd859fd2c73c4862808424b3df6a1e27e7],Size_:374119253,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=0a9ef099-e20e-4fe4-96ed-6e360c5e8cec name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.151982984Z" level=info msg="Creating container: openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-pqwkf/prometheus-operator-admission-webhook" id=94b9641c-783a-44a1-ab34-4d05b16a507b name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.152110876Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.154759676Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7" id=1ef03c1f-511b-4058-9843-6b9232e8354e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.198854733Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:4cb41c73a232ec4401c5db32cdb523810dc12ac605cdc6a9ce388e90930eb6c4,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:304fb0634e581d983684570720cd214339fbb3a85299cb5d1e1470e8946c52b7],Size_:515955107,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=1ef03c1f-511b-4058-9843-6b9232e8354e name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.200980456Z" level=info msg="Creating container: openshift-network-diagnostics/network-check-source-6775957967-7tm54/check-endpoints" id=5f3b6992-5e74-42fa-bc09-02a758dad506 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.201079504Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.339647 1977 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-ingress/router-default-5cf6bbdff9-4vdhd] Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.346577600Z" level=info msg="Created container 022fbbd407117de7d5e65652df1c7e0857327f32f625b473dd9ff5b14d42991a: openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc/kube-rbac-proxy" id=b4ebb602-5c71-4bb4-9e86-f55bf115043f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.350233627Z" level=info msg="Starting container: 022fbbd407117de7d5e65652df1c7e0857327f32f625b473dd9ff5b14d42991a" id=2f2a757e-44cf-4366-b666-5c0725b4b191 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: 2023-01-05T09:47:07Z [verbose] Add: openshift-ingress:router-default-5cf6bbdff9-4vdhd:78fead4b-9b2f-4ed6-a4f7-4c8479e076ee:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/dbfdeaf8-528d-4e2e-b7d2-9087124a29f7"}],"ips":[{"version":"4","interface":0,"address":"10.131.0.10/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: I0105 09:47:07.290325 4759 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-ingress", Name:"router-default-5cf6bbdff9-4vdhd", UID:"78fead4b-9b2f-4ed6-a4f7-4c8479e076ee", APIVersion:"v1", ResourceVersion:"107764", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.10/23] from openshift-sdn Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.352907865Z" level=info msg="Got pod network &{Name:router-default-5cf6bbdff9-4vdhd Namespace:openshift-ingress ID:4660d5572cae6cef99bbb174a92ded51ecce6bfa33a8aede8e8d313ee080384d UID:78fead4b-9b2f-4ed6-a4f7-4c8479e076ee NetNS:/var/run/netns/dbfdeaf8-528d-4e2e-b7d2-9087124a29f7 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.353082951Z" level=info msg="Checking pod openshift-ingress_router-default-5cf6bbdff9-4vdhd for CNI network multus-cni-network (type=multus)" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: W0105 09:47:07.366415 1977 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod78fead4b_9b2f_4ed6_a4f7_4c8479e076ee.slice/crio-4660d5572cae6cef99bbb174a92ded51ecce6bfa33a8aede8e8d313ee080384d.scope WatchSource:0}: Error finding container 4660d5572cae6cef99bbb174a92ded51ecce6bfa33a8aede8e8d313ee080384d: Status 404 returned error can't find the container with id 4660d5572cae6cef99bbb174a92ded51ecce6bfa33a8aede8e8d313ee080384d Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.384870276Z" level=info msg="Ran pod sandbox 4660d5572cae6cef99bbb174a92ded51ecce6bfa33a8aede8e8d313ee080384d with infra container: openshift-ingress/router-default-5cf6bbdff9-4vdhd/POD" id=e4df48a1-da4f-46af-9abd-0758b8d66857 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.401205351Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:bfd14188641a2d34e5b2b2ed3336b3b440155533f3fd7c6de776300dc3ff9e3b" id=979c1119-2667-4e5a-8c52-d367fe418c01 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.401425981Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:ad00ebc82f5cc671785b8401477b42db226d1c335ea60a4bb39bf28a961de840,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:bfd14188641a2d34e5b2b2ed3336b3b440155533f3fd7c6de776300dc3ff9e3b],Size_:411774965,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=979c1119-2667-4e5a-8c52-d367fe418c01 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.403792 1977 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-monitoring/alertmanager-main-1] Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.403847 1977 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.408667034Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:bfd14188641a2d34e5b2b2ed3336b3b440155533f3fd7c6de776300dc3ff9e3b" id=96f7310e-904d-4b58-9939-7342e0001e77 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.408885301Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:ad00ebc82f5cc671785b8401477b42db226d1c335ea60a4bb39bf28a961de840,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:bfd14188641a2d34e5b2b2ed3336b3b440155533f3fd7c6de776300dc3ff9e3b],Size_:411774965,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=96f7310e-904d-4b58-9939-7342e0001e77 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.409210068Z" level=info msg="Started container" PID=5090 containerID=022fbbd407117de7d5e65652df1c7e0857327f32f625b473dd9ff5b14d42991a description=openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc/kube-rbac-proxy id=2f2a757e-44cf-4366-b666-5c0725b4b191 name=/runtime.v1.RuntimeService/StartContainer sandboxID=54dbaba32104ac88a1e3ed7c1c5c8cd5be28a9c930620bf99f4e31003ced75b4 Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.412816995Z" level=info msg="Creating container: openshift-ingress/router-default-5cf6bbdff9-4vdhd/router" id=42e50b3c-c7b8-4544-aeca-8c6d00e3d710 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.412987745Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.474536479Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816" id=a3c58ac7-992b-4324-9d68-0106a5a8cb9c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.504283 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jf57m\" (UniqueName: \"kubernetes.io/projected/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-kube-api-access-jf57m\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.504353 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-tls-assets\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.504387 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-metrics-client-ca\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.504438 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-config-out\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.504473 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-secret-alertmanager-main-tls\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.504510 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.504551 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-alertmanager-main-db\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.504586 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-config-volume\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.504621 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-proxy\" (UniqueName: \"kubernetes.io/secret/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-secret-alertmanager-main-proxy\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.504659 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.504709 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.504743 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-web-config\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.546473505Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:955ee2b41ecd88e3a9c0183a0b65d6edbf524cefa1318837dcbc82b0466c4d55,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816],Size_:309576938,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=a3c58ac7-992b-4324-9d68-0106a5a8cb9c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.547762968Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816" id=2ed768ef-5448-46bf-8460-c645676e5f18 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.548007121Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:955ee2b41ecd88e3a9c0183a0b65d6edbf524cefa1318837dcbc82b0466c4d55,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816],Size_:309576938,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=2ed768ef-5448-46bf-8460-c645676e5f18 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.549125669Z" level=info msg="Creating container: openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc/prom-label-proxy" id=8aa5e7ee-5869-42df-b6f0-32f6357ec714 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.549269406Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.624922 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-web-config\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.624998 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-jf57m\" (UniqueName: \"kubernetes.io/projected/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-kube-api-access-jf57m\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.625041 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-tls-assets\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.625079 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-metrics-client-ca\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.625123 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-config-out\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.625161 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-secret-alertmanager-main-tls\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.625199 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.625233 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-alertmanager-main-db\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.625271 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-config-volume\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.625308 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-proxy\" (UniqueName: \"kubernetes.io/secret/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-secret-alertmanager-main-proxy\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.625343 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.625391 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.627967 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.630214 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-web-config\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.637398 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.637975 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-metrics-client-ca\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.640006 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-config-out\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.640511 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-tls-assets\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.640731 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-alertmanager-main-db\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.646514 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-config-volume\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.646926 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-secret-alertmanager-main-tls\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.654282 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-proxy\" (UniqueName: \"kubernetes.io/secret/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-secret-alertmanager-main-proxy\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.658028 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.668060 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-pqwkf" event=&{ID:f7f51e23-4c42-4696-b073-e60655077a47 Type:ContainerStarted Data:31719a55b8c87ba34a4133b598d3eb7b108f583c8f149bd000793f3743c53086} Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.669463 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" event=&{ID:3b3a8ee0-19b5-43a0-b793-cb6e390cba80 Type:ContainerStarted Data:022fbbd407117de7d5e65652df1c7e0857327f32f625b473dd9ff5b14d42991a} Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.669487 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" event=&{ID:3b3a8ee0-19b5-43a0-b793-cb6e390cba80 Type:ContainerStarted Data:d84da8bec40a36151af3705b3b83dd377e211227e824a7ad34e3a18086ec78d8} Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.670231 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-6775957967-7tm54" event=&{ID:85256957-515d-46bc-801a-4024088469ed Type:ContainerStarted Data:dd5d6f1a8a242dced9e946b55763f6bdd8ee37d23cb679a2b3b0dde79120ed24} Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.671282 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2" event=&{ID:a6b28bc9-5526-449d-b81c-ba4c654e254b Type:ContainerStarted Data:846ffcc9f1e0feb970619b56fe60c2f15822dda468bcd0552ca0a3625b6c58b7} Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.671305 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2" event=&{ID:a6b28bc9-5526-449d-b81c-ba4c654e254b Type:ContainerStarted Data:ad84fb5151c7b562c151f9b22fc3980b8cb0e4f105022cec531ada830a445264} Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.672368 1977 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.673084 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5cf6bbdff9-4vdhd" event=&{ID:78fead4b-9b2f-4ed6-a4f7-4c8479e076ee Type:ContainerStarted Data:4660d5572cae6cef99bbb174a92ded51ecce6bfa33a8aede8e8d313ee080384d} Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.711509 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-jf57m\" (UniqueName: \"kubernetes.io/projected/d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46-kube-api-access-jf57m\") pod \"alertmanager-main-1\" (UID: \"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46\") " pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.738659 1977 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-monitoring/prometheus-k8s-1] Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.738704 1977 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.810759 1977 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.821091744Z" level=info msg="Running pod sandbox: openshift-monitoring/alertmanager-main-1/POD" id=018eec46-9117-48af-9356-a0686a5126ad name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.821149044Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.938178 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e5650cc7-8fac-496c-8567-d78340f40f5e-config\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.938235 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/e5650cc7-8fac-496c-8567-d78340f40f5e-prometheus-k8s-db\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.938271 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/e5650cc7-8fac-496c-8567-d78340f40f5e-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.938304 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/e5650cc7-8fac-496c-8567-d78340f40f5e-configmap-metrics-client-ca\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.938336 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/e5650cc7-8fac-496c-8567-d78340f40f5e-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.938322814Z" level=info msg="Created container 3331f4bea8df82707cd4202fb0d45f7a2fb3234e8d3b191046d119ed966f5ad9: openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-pqwkf/prometheus-operator-admission-webhook" id=94b9641c-783a-44a1-ab34-4d05b16a507b name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.938374 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e5650cc7-8fac-496c-8567-d78340f40f5e-tls-assets\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.938408 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5650cc7-8fac-496c-8567-d78340f40f5e-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.938445 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-kube-etcd-client-certs\" (UniqueName: \"kubernetes.io/secret/e5650cc7-8fac-496c-8567-d78340f40f5e-secret-kube-etcd-client-certs\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.938478 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-proxy\" (UniqueName: \"kubernetes.io/secret/e5650cc7-8fac-496c-8567-d78340f40f5e-secret-prometheus-k8s-proxy\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.938515 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e5650cc7-8fac-496c-8567-d78340f40f5e-web-config\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.938545 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jdxt\" (UniqueName: \"kubernetes.io/projected/e5650cc7-8fac-496c-8567-d78340f40f5e-kube-api-access-2jdxt\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.938575 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/e5650cc7-8fac-496c-8567-d78340f40f5e-secret-grpc-tls\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.938607 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/e5650cc7-8fac-496c-8567-d78340f40f5e-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.938647 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/e5650cc7-8fac-496c-8567-d78340f40f5e-metrics-client-ca\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.938689 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5650cc7-8fac-496c-8567-d78340f40f5e-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.938724 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/e5650cc7-8fac-496c-8567-d78340f40f5e-secret-metrics-client-certs\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.938761 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5650cc7-8fac-496c-8567-d78340f40f5e-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.938796 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e5650cc7-8fac-496c-8567-d78340f40f5e-config-out\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:07.938828 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/e5650cc7-8fac-496c-8567-d78340f40f5e-secret-kube-rbac-proxy\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.940272724Z" level=info msg="Starting container: 3331f4bea8df82707cd4202fb0d45f7a2fb3234e8d3b191046d119ed966f5ad9" id=6c8cc90a-6658-42a5-ba72-d01e89b16646 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.953318246Z" level=info msg="Got pod network &{Name:alertmanager-main-1 Namespace:openshift-monitoring ID:6f2169b19a28ba86c3b28c7e7891c0a5cc09b988030c31c54bd83e2c69110f12 UID:d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46 NetNS:/var/run/netns/c5bf21f0-80b3-4b4c-9279-af2a98879da2 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:47:07 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:07.953351679Z" level=info msg="Adding pod openshift-monitoring_alertmanager-main-1 to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:08.015869187Z" level=info msg="Started container" PID=5156 containerID=3331f4bea8df82707cd4202fb0d45f7a2fb3234e8d3b191046d119ed966f5ad9 description=openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-pqwkf/prometheus-operator-admission-webhook id=6c8cc90a-6658-42a5-ba72-d01e89b16646 name=/runtime.v1.RuntimeService/StartContainer sandboxID=31719a55b8c87ba34a4133b598d3eb7b108f583c8f149bd000793f3743c53086 Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.058446 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5650cc7-8fac-496c-8567-d78340f40f5e-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.058501 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-kube-etcd-client-certs\" (UniqueName: \"kubernetes.io/secret/e5650cc7-8fac-496c-8567-d78340f40f5e-secret-kube-etcd-client-certs\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.058536 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-proxy\" (UniqueName: \"kubernetes.io/secret/e5650cc7-8fac-496c-8567-d78340f40f5e-secret-prometheus-k8s-proxy\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.058576 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-2jdxt\" (UniqueName: \"kubernetes.io/projected/e5650cc7-8fac-496c-8567-d78340f40f5e-kube-api-access-2jdxt\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.058619 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e5650cc7-8fac-496c-8567-d78340f40f5e-web-config\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.058656 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/e5650cc7-8fac-496c-8567-d78340f40f5e-secret-grpc-tls\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.058694 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/e5650cc7-8fac-496c-8567-d78340f40f5e-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.058725 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/e5650cc7-8fac-496c-8567-d78340f40f5e-metrics-client-ca\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.058758 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5650cc7-8fac-496c-8567-d78340f40f5e-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.058787 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/e5650cc7-8fac-496c-8567-d78340f40f5e-secret-metrics-client-certs\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.058820 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e5650cc7-8fac-496c-8567-d78340f40f5e-config-out\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.058868 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5650cc7-8fac-496c-8567-d78340f40f5e-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.058910 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/e5650cc7-8fac-496c-8567-d78340f40f5e-secret-kube-rbac-proxy\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.058968 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e5650cc7-8fac-496c-8567-d78340f40f5e-config\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.059006 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/e5650cc7-8fac-496c-8567-d78340f40f5e-prometheus-k8s-db\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.059044 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/e5650cc7-8fac-496c-8567-d78340f40f5e-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.059078 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/e5650cc7-8fac-496c-8567-d78340f40f5e-configmap-metrics-client-ca\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.059118 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/e5650cc7-8fac-496c-8567-d78340f40f5e-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.059165 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e5650cc7-8fac-496c-8567-d78340f40f5e-tls-assets\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.060288 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5650cc7-8fac-496c-8567-d78340f40f5e-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.061044 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5650cc7-8fac-496c-8567-d78340f40f5e-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.067126 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e5650cc7-8fac-496c-8567-d78340f40f5e-tls-assets\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.069480 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/e5650cc7-8fac-496c-8567-d78340f40f5e-secret-metrics-client-certs\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.069767 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e5650cc7-8fac-496c-8567-d78340f40f5e-config-out\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.070438 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5650cc7-8fac-496c-8567-d78340f40f5e-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:08.082603099Z" level=info msg="Created container 87b293dc9f89b413924acdf263e63f65ebd3a35b28bdd1149d01f10da98f5809: openshift-network-diagnostics/network-check-source-6775957967-7tm54/check-endpoints" id=5f3b6992-5e74-42fa-bc09-02a758dad506 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.085612 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/e5650cc7-8fac-496c-8567-d78340f40f5e-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.089966 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/e5650cc7-8fac-496c-8567-d78340f40f5e-prometheus-k8s-db\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.112624 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/e5650cc7-8fac-496c-8567-d78340f40f5e-metrics-client-ca\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.114607 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/e5650cc7-8fac-496c-8567-d78340f40f5e-configmap-metrics-client-ca\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.128990 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/e5650cc7-8fac-496c-8567-d78340f40f5e-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:08.129512212Z" level=info msg="Starting container: 87b293dc9f89b413924acdf263e63f65ebd3a35b28bdd1149d01f10da98f5809" id=66806193-b190-41b8-8e5a-aed08f7c86f4 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.134400 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/e5650cc7-8fac-496c-8567-d78340f40f5e-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.142637 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jdxt\" (UniqueName: \"kubernetes.io/projected/e5650cc7-8fac-496c-8567-d78340f40f5e-kube-api-access-2jdxt\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.143041 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e5650cc7-8fac-496c-8567-d78340f40f5e-config\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.143420 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-proxy\" (UniqueName: \"kubernetes.io/secret/e5650cc7-8fac-496c-8567-d78340f40f5e-secret-prometheus-k8s-proxy\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.152010 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-kube-etcd-client-certs\" (UniqueName: \"kubernetes.io/secret/e5650cc7-8fac-496c-8567-d78340f40f5e-secret-kube-etcd-client-certs\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:08.159852990Z" level=info msg="Started container" PID=5190 containerID=87b293dc9f89b413924acdf263e63f65ebd3a35b28bdd1149d01f10da98f5809 description=openshift-network-diagnostics/network-check-source-6775957967-7tm54/check-endpoints id=66806193-b190-41b8-8e5a-aed08f7c86f4 name=/runtime.v1.RuntimeService/StartContainer sandboxID=dd5d6f1a8a242dced9e946b55763f6bdd8ee37d23cb679a2b3b0dde79120ed24 Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.177727 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/e5650cc7-8fac-496c-8567-d78340f40f5e-secret-kube-rbac-proxy\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.178856 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/e5650cc7-8fac-496c-8567-d78340f40f5e-secret-grpc-tls\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.180301 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e5650cc7-8fac-496c-8567-d78340f40f5e-web-config\") pod \"prometheus-k8s-1\" (UID: \"e5650cc7-8fac-496c-8567-d78340f40f5e\") " pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:08.311938306Z" level=info msg="Running pod sandbox: openshift-monitoring/prometheus-k8s-1/POD" id=92e06592-acad-4dc9-be31-b139de5bcf12 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:08.312032155Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.311279 1977 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:08.400005908Z" level=info msg="Created container f3694a10e7bb76fd98830318febd5c83fdd468134330f44c4a0efc3aa49c03a7: openshift-ingress/router-default-5cf6bbdff9-4vdhd/router" id=42e50b3c-c7b8-4544-aeca-8c6d00e3d710 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:08.411212357Z" level=info msg="Starting container: f3694a10e7bb76fd98830318febd5c83fdd468134330f44c4a0efc3aa49c03a7" id=68521778-e40e-4883-aa0f-b4508f16eecc name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:08.412974214Z" level=info msg="Got pod network &{Name:prometheus-k8s-1 Namespace:openshift-monitoring ID:1df490500596677201ef8a2dc1ead29190963e5f81f951619c88d10237271737 UID:e5650cc7-8fac-496c-8567-d78340f40f5e NetNS:/var/run/netns/b28a0c3c-c56b-4eab-bf62-bbfe51284fd0 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:08.413004105Z" level=info msg="Adding pod openshift-monitoring_prometheus-k8s-1 to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:08.504032237Z" level=info msg="Started container" PID=5324 containerID=f3694a10e7bb76fd98830318febd5c83fdd468134330f44c4a0efc3aa49c03a7 description=openshift-ingress/router-default-5cf6bbdff9-4vdhd/router id=68521778-e40e-4883-aa0f-b4508f16eecc name=/runtime.v1.RuntimeService/StartContainer sandboxID=4660d5572cae6cef99bbb174a92ded51ecce6bfa33a8aede8e8d313ee080384d Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:08.569557320Z" level=info msg="Created container d66b7283c113ae20e6bd7c7516351ae53cb2256b20d9a4cc0864628eb5172bae: openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc/prom-label-proxy" id=8aa5e7ee-5869-42df-b6f0-32f6357ec714 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:08.570133252Z" level=info msg="Starting container: d66b7283c113ae20e6bd7c7516351ae53cb2256b20d9a4cc0864628eb5172bae" id=9bebd2bc-947f-4ca9-9513-f177fd3c5c35 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:08.645387275Z" level=info msg="Started container" PID=5283 containerID=d66b7283c113ae20e6bd7c7516351ae53cb2256b20d9a4cc0864628eb5172bae description=openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc/prom-label-proxy id=9bebd2bc-947f-4ca9-9513-f177fd3c5c35 name=/runtime.v1.RuntimeService/StartContainer sandboxID=54dbaba32104ac88a1e3ed7c1c5c8cd5be28a9c930620bf99f4e31003ced75b4 Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:08.665679067Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=f8443796-2a5f-42ce-9c6c-d9c281f72d58 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:08.665921402Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=f8443796-2a5f-42ce-9c6c-d9c281f72d58 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:08.668883327Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=fdc3188d-abea-4562-8f94-e5690fc47d4c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:08.669099333Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=fdc3188d-abea-4562-8f94-e5690fc47d4c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:08.670241273Z" level=info msg="Creating container: openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc/kube-rbac-proxy-rules" id=2276c228-a660-46a0-bd28-ac1138fb0ac6 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:08.670362457Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.680128 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-6775957967-7tm54" event=&{ID:85256957-515d-46bc-801a-4024088469ed Type:ContainerStarted Data:87b293dc9f89b413924acdf263e63f65ebd3a35b28bdd1149d01f10da98f5809} Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.682488 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5cf6bbdff9-4vdhd" event=&{ID:78fead4b-9b2f-4ed6-a4f7-4c8479e076ee Type:ContainerStarted Data:f3694a10e7bb76fd98830318febd5c83fdd468134330f44c4a0efc3aa49c03a7} Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.690478 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-pqwkf" event=&{ID:f7f51e23-4c42-4696-b073-e60655077a47 Type:ContainerStarted Data:3331f4bea8df82707cd4202fb0d45f7a2fb3234e8d3b191046d119ed966f5ad9} Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.691555 1977 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-pqwkf" Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.695799 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" event=&{ID:3b3a8ee0-19b5-43a0-b793-cb6e390cba80 Type:ContainerStarted Data:d66b7283c113ae20e6bd7c7516351ae53cb2256b20d9a4cc0864628eb5172bae} Jan 05 09:47:08 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:08.725383 1977 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-598468bb49-pqwkf" Jan 05 09:47:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:09.033248 1977 kubelet.go:2229] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5cf6bbdff9-4vdhd" Jan 05 09:47:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:09.071035 1977 patch_prober.go:29] interesting pod/router-default-5cf6bbdff9-4vdhd container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 05 09:47:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: [+]has-synced ok Jan 05 09:47:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: [+]process-running ok Jan 05 09:47:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: healthz check failed Jan 05 09:47:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:09.071099 1977 prober.go:114] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5cf6bbdff9-4vdhd" podUID=78fead4b-9b2f-4ed6-a4f7-4c8479e076ee containerName="router" probeResult=failure output="HTTP probe failed with statuscode: 500" Jan 05 09:47:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:09.216222634Z" level=info msg="Created container 279755d6781821f9cf20ad397b3e948b45312e66b52ac5d95f18062d5037484a: openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc/kube-rbac-proxy-rules" id=2276c228-a660-46a0-bd28-ac1138fb0ac6 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:09.216875343Z" level=info msg="Starting container: 279755d6781821f9cf20ad397b3e948b45312e66b52ac5d95f18062d5037484a" id=795f63fd-975d-42fb-8b10-2b0635462e58 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:47:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:09.258481391Z" level=info msg="Started container" PID=5467 containerID=279755d6781821f9cf20ad397b3e948b45312e66b52ac5d95f18062d5037484a description=openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc/kube-rbac-proxy-rules id=795f63fd-975d-42fb-8b10-2b0635462e58 name=/runtime.v1.RuntimeService/StartContainer sandboxID=54dbaba32104ac88a1e3ed7c1c5c8cd5be28a9c930620bf99f4e31003ced75b4 Jan 05 09:47:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:09.290819455Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=0367bec6-8d75-49e7-9909-0136b74f975d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:09.291064095Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=0367bec6-8d75-49e7-9909-0136b74f975d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:09.292075213Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=a77b24df-c7bf-4794-b456-7cac8a108f8b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:09.292260230Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=a77b24df-c7bf-4794-b456-7cac8a108f8b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:09.293218559Z" level=info msg="Creating container: openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc/kube-rbac-proxy-metrics" id=38bf8f6d-5814-44ca-9ca5-997e4a742e01 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:09.293334771Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:47:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:09.504667034Z" level=info msg="Created container 13401b4730ccc663d1124e4f1ed5c21ef935a2516da02e7775445c9f2c79d2e2: openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc/kube-rbac-proxy-metrics" id=38bf8f6d-5814-44ca-9ca5-997e4a742e01 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:09.505139653Z" level=info msg="Starting container: 13401b4730ccc663d1124e4f1ed5c21ef935a2516da02e7775445c9f2c79d2e2" id=bc560e93-b753-4e39-bca5-2bf03fd704e3 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:47:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:09.530025953Z" level=info msg="Started container" PID=5515 containerID=13401b4730ccc663d1124e4f1ed5c21ef935a2516da02e7775445c9f2c79d2e2 description=openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc/kube-rbac-proxy-metrics id=bc560e93-b753-4e39-bca5-2bf03fd704e3 name=/runtime.v1.RuntimeService/StartContainer sandboxID=54dbaba32104ac88a1e3ed7c1c5c8cd5be28a9c930620bf99f4e31003ced75b4 Jan 05 09:47:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:09.701361 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" event=&{ID:3b3a8ee0-19b5-43a0-b793-cb6e390cba80 Type:ContainerStarted Data:13401b4730ccc663d1124e4f1ed5c21ef935a2516da02e7775445c9f2c79d2e2} Jan 05 09:47:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:09.701404 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" event=&{ID:3b3a8ee0-19b5-43a0-b793-cb6e390cba80 Type:ContainerStarted Data:279755d6781821f9cf20ad397b3e948b45312e66b52ac5d95f18062d5037484a} Jan 05 09:47:09 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:09.703779 1977 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:10.017440 1977 kubelet.go:2229] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5cf6bbdff9-4vdhd" Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:10.073530 1977 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/prometheus-k8s-1] Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: 2023-01-05T09:47:10Z [verbose] Add: openshift-monitoring:prometheus-k8s-1:e5650cc7-8fac-496c-8567-d78340f40f5e:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/b28a0c3c-c56b-4eab-bf62-bbfe51284fd0"}],"ips":[{"version":"4","interface":0,"address":"10.131.0.12/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: I0105 09:47:10.047990 5349 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"prometheus-k8s-1", UID:"e5650cc7-8fac-496c-8567-d78340f40f5e", APIVersion:"v1", ResourceVersion:"107900", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.12/23] from openshift-sdn Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.078500387Z" level=info msg="Got pod network &{Name:prometheus-k8s-1 Namespace:openshift-monitoring ID:1df490500596677201ef8a2dc1ead29190963e5f81f951619c88d10237271737 UID:e5650cc7-8fac-496c-8567-d78340f40f5e NetNS:/var/run/netns/b28a0c3c-c56b-4eab-bf62-bbfe51284fd0 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.078662815Z" level=info msg="Checking pod openshift-monitoring_prometheus-k8s-1 for CNI network multus-cni-network (type=multus)" Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.098459349Z" level=info msg="Ran pod sandbox 1df490500596677201ef8a2dc1ead29190963e5f81f951619c88d10237271737 with infra container: openshift-monitoring/prometheus-k8s-1/POD" id=92e06592-acad-4dc9-be31-b139de5bcf12 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.099365372Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=ec248d16-8c37-4ad2-85e0-b05d5b48b20f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.099572097Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=ec248d16-8c37-4ad2-85e0-b05d5b48b20f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.100305717Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=23717c7d-89c2-45bc-a2b0-97844777ac2c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.100510749Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=23717c7d-89c2-45bc-a2b0-97844777ac2c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.102568752Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-1/init-config-reloader" id=e57328a5-4c9c-48ed-879d-f4299ce9c3ea name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.102684521Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:10.181789 1977 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-monitoring/alertmanager-main-1] Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: 2023-01-05T09:47:10Z [verbose] Add: openshift-monitoring:alertmanager-main-1:d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/c5bf21f0-80b3-4b4c-9279-af2a98879da2"}],"ips":[{"version":"4","interface":0,"address":"10.131.0.11/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: I0105 09:47:10.144809 5254 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-monitoring", Name:"alertmanager-main-1", UID:"d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46", APIVersion:"v1", ResourceVersion:"107868", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.11/23] from openshift-sdn Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.191373133Z" level=info msg="Got pod network &{Name:alertmanager-main-1 Namespace:openshift-monitoring ID:6f2169b19a28ba86c3b28c7e7891c0a5cc09b988030c31c54bd83e2c69110f12 UID:d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46 NetNS:/var/run/netns/c5bf21f0-80b3-4b4c-9279-af2a98879da2 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.191513505Z" level=info msg="Checking pod openshift-monitoring_alertmanager-main-1 for CNI network multus-cni-network (type=multus)" Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.211865940Z" level=info msg="Ran pod sandbox 6f2169b19a28ba86c3b28c7e7891c0a5cc09b988030c31c54bd83e2c69110f12 with infra container: openshift-monitoring/alertmanager-main-1/POD" id=018eec46-9117-48af-9356-a0686a5126ad name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: W0105 09:47:10.219350 1977 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd37fa6e9_f7ae_46f4_bf1b_d6e57f095d46.slice/crio-6f2169b19a28ba86c3b28c7e7891c0a5cc09b988030c31c54bd83e2c69110f12.scope WatchSource:0}: Error finding container 6f2169b19a28ba86c3b28c7e7891c0a5cc09b988030c31c54bd83e2c69110f12: Status 404 returned error can't find the container with id 6f2169b19a28ba86c3b28c7e7891c0a5cc09b988030c31c54bd83e2c69110f12 Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.229613651Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e6173b5d77def9e8058c7ff82f1ad65f537d1a81f88fcdabe6d8e30bff49ca89" id=281d199d-4ce5-412c-b8a1-487aa0e64b2c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.229819761Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:d168c738461c25cc0c3fa96cfcdf2dc337ca08f5bf3a6b550fcb29a59ea39270,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e6173b5d77def9e8058c7ff82f1ad65f537d1a81f88fcdabe6d8e30bff49ca89],Size_:333780768,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=281d199d-4ce5-412c-b8a1-487aa0e64b2c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.266372301Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e6173b5d77def9e8058c7ff82f1ad65f537d1a81f88fcdabe6d8e30bff49ca89" id=a2eb1ee9-a8b8-4fea-936b-a1f0c40fe2a8 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.266578195Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:d168c738461c25cc0c3fa96cfcdf2dc337ca08f5bf3a6b550fcb29a59ea39270,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:e6173b5d77def9e8058c7ff82f1ad65f537d1a81f88fcdabe6d8e30bff49ca89],Size_:333780768,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=a2eb1ee9-a8b8-4fea-936b-a1f0c40fe2a8 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.272088722Z" level=info msg="Creating container: openshift-monitoring/alertmanager-main-1/alertmanager" id=099d9f79-8ebb-472c-99dd-24de62c6b0a8 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.272217734Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.441723588Z" level=info msg="Created container deccfe5f6565b8bdec56c8a4f9fe0c83397ef94cd2a0102b2148e6ce64a42b02: openshift-monitoring/prometheus-k8s-1/init-config-reloader" id=e57328a5-4c9c-48ed-879d-f4299ce9c3ea name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.448204494Z" level=info msg="Starting container: deccfe5f6565b8bdec56c8a4f9fe0c83397ef94cd2a0102b2148e6ce64a42b02" id=e2f381ed-81bb-483b-ae9a-5ddd7e17d4bb name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.448804579Z" level=info msg="Created container f42f3cb33a6a9c4718fc2487a56da85dbfb7923e0e443002774fbf1a176a4658: openshift-monitoring/alertmanager-main-1/alertmanager" id=099d9f79-8ebb-472c-99dd-24de62c6b0a8 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.449226244Z" level=info msg="Starting container: f42f3cb33a6a9c4718fc2487a56da85dbfb7923e0e443002774fbf1a176a4658" id=d21b450d-f29b-4be6-a812-0bb1d6679e45 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.480048966Z" level=info msg="Started container" PID=5607 containerID=deccfe5f6565b8bdec56c8a4f9fe0c83397ef94cd2a0102b2148e6ce64a42b02 description=openshift-monitoring/prometheus-k8s-1/init-config-reloader id=e2f381ed-81bb-483b-ae9a-5ddd7e17d4bb name=/runtime.v1.RuntimeService/StartContainer sandboxID=1df490500596677201ef8a2dc1ead29190963e5f81f951619c88d10237271737 Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.491139030Z" level=info msg="Started container" PID=5623 containerID=f42f3cb33a6a9c4718fc2487a56da85dbfb7923e0e443002774fbf1a176a4658 description=openshift-monitoring/alertmanager-main-1/alertmanager id=d21b450d-f29b-4be6-a812-0bb1d6679e45 name=/runtime.v1.RuntimeService/StartContainer sandboxID=6f2169b19a28ba86c3b28c7e7891c0a5cc09b988030c31c54bd83e2c69110f12 Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.515801520Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=facb06b6-93ee-4de2-8e5e-ee922bbdb74a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.516045302Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=facb06b6-93ee-4de2-8e5e-ee922bbdb74a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.519206210Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=bc708152-7b54-4268-acaf-f9d9f60baa36 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.519395173Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=bc708152-7b54-4268-acaf-f9d9f60baa36 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.520549724Z" level=info msg="Creating container: openshift-monitoring/alertmanager-main-1/config-reloader" id=80582c97-974e-42e6-8b69-fadba8529a8d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.520661811Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:10.707154 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-1" event=&{ID:d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46 Type:ContainerStarted Data:f42f3cb33a6a9c4718fc2487a56da85dbfb7923e0e443002774fbf1a176a4658} Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:10.707189 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-1" event=&{ID:d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46 Type:ContainerStarted Data:6f2169b19a28ba86c3b28c7e7891c0a5cc09b988030c31c54bd83e2c69110f12} Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:10.710007 1977 generic.go:296] "Generic (PLEG): container finished" podID=e5650cc7-8fac-496c-8567-d78340f40f5e containerID="deccfe5f6565b8bdec56c8a4f9fe0c83397ef94cd2a0102b2148e6ce64a42b02" exitCode=0 Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:10.711268 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-1" event=&{ID:e5650cc7-8fac-496c-8567-d78340f40f5e Type:ContainerDied Data:deccfe5f6565b8bdec56c8a4f9fe0c83397ef94cd2a0102b2148e6ce64a42b02} Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:10.711297 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-1" event=&{ID:e5650cc7-8fac-496c-8567-d78340f40f5e Type:ContainerStarted Data:1df490500596677201ef8a2dc1ead29190963e5f81f951619c88d10237271737} Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:10.714414 1977 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5cf6bbdff9-4vdhd" Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.714814524Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:91e7d68b28aa00e209065604b2b9ac1400671d9102581e8b49b37726acd7b6e4" id=870738d9-8084-4714-806b-1d181bf92a6c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.715078154Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b68b0fa7ea4653b56a91a8c213d456cab2f645207af609598f9877a41585eeae,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:91e7d68b28aa00e209065604b2b9ac1400671d9102581e8b49b37726acd7b6e4],Size_:493220384,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=870738d9-8084-4714-806b-1d181bf92a6c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:10.717111 1977 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5cf6bbdff9-4vdhd" Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.726185370Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:91e7d68b28aa00e209065604b2b9ac1400671d9102581e8b49b37726acd7b6e4" id=4b84a273-dd38-4e4e-9b93-8cd81e0aa6ef name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.726383648Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:b68b0fa7ea4653b56a91a8c213d456cab2f645207af609598f9877a41585eeae,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:91e7d68b28aa00e209065604b2b9ac1400671d9102581e8b49b37726acd7b6e4],Size_:493220384,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=4b84a273-dd38-4e4e-9b93-8cd81e0aa6ef name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.727881245Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-1/prometheus" id=0c54e8f5-ae4c-4ed9-bbe5-e5e005f87ec2 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.728015875Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.815928506Z" level=info msg="Created container 68d3d2638852756ec6ba856e2401d3f254fdd1ac1a0e149de03ac08298f12bc1: openshift-monitoring/alertmanager-main-1/config-reloader" id=80582c97-974e-42e6-8b69-fadba8529a8d name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.823090363Z" level=info msg="Starting container: 68d3d2638852756ec6ba856e2401d3f254fdd1ac1a0e149de03ac08298f12bc1" id=04cbd426-0540-4a7a-8cdd-4d5b1ea0b12b name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.842032652Z" level=info msg="Started container" PID=5688 containerID=68d3d2638852756ec6ba856e2401d3f254fdd1ac1a0e149de03ac08298f12bc1 description=openshift-monitoring/alertmanager-main-1/config-reloader id=04cbd426-0540-4a7a-8cdd-4d5b1ea0b12b name=/runtime.v1.RuntimeService/StartContainer sandboxID=6f2169b19a28ba86c3b28c7e7891c0a5cc09b988030c31c54bd83e2c69110f12 Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.873069988Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=70d13174-6611-46c2-8a80-5005610bc8d7 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.873312312Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=70d13174-6611-46c2-8a80-5005610bc8d7 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.874353795Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=25aac3c9-08e7-4a2c-88b6-b489dac78dbe name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.874544273Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=25aac3c9-08e7-4a2c-88b6-b489dac78dbe name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.875852765Z" level=info msg="Creating container: openshift-monitoring/alertmanager-main-1/alertmanager-proxy" id=2bf6d3a6-ad2d-4a60-986e-183e486d3e10 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:10 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:10.875997169Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.107223200Z" level=info msg="Created container 9770e5949056ad84e03a67e5dfc68ea28b04d2c3058d8f91d131ef605477df7b: openshift-monitoring/prometheus-k8s-1/prometheus" id=0c54e8f5-ae4c-4ed9-bbe5-e5e005f87ec2 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.113056262Z" level=info msg="Starting container: 9770e5949056ad84e03a67e5dfc68ea28b04d2c3058d8f91d131ef605477df7b" id=97e63ee2-d8ba-4bff-b832-fd81b520330e name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.129221622Z" level=info msg="Started container" PID=5753 containerID=9770e5949056ad84e03a67e5dfc68ea28b04d2c3058d8f91d131ef605477df7b description=openshift-monitoring/prometheus-k8s-1/prometheus id=97e63ee2-d8ba-4bff-b832-fd81b520330e name=/runtime.v1.RuntimeService/StartContainer sandboxID=1df490500596677201ef8a2dc1ead29190963e5f81f951619c88d10237271737 Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.139674499Z" level=info msg="Created container d3708188be0373e12e4347439a1a6bf7c8fa134f158a90035cbe341294b21505: openshift-monitoring/alertmanager-main-1/alertmanager-proxy" id=2bf6d3a6-ad2d-4a60-986e-183e486d3e10 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.140184872Z" level=info msg="Starting container: d3708188be0373e12e4347439a1a6bf7c8fa134f158a90035cbe341294b21505" id=f903d2f0-814e-4462-aa04-2edc8f2cac78 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.176612878Z" level=info msg="Started container" PID=5776 containerID=d3708188be0373e12e4347439a1a6bf7c8fa134f158a90035cbe341294b21505 description=openshift-monitoring/alertmanager-main-1/alertmanager-proxy id=f903d2f0-814e-4462-aa04-2edc8f2cac78 name=/runtime.v1.RuntimeService/StartContainer sandboxID=6f2169b19a28ba86c3b28c7e7891c0a5cc09b988030c31c54bd83e2c69110f12 Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.205010270Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=8d22d4d9-c241-4b10-a75e-d7c09c3d4ecb name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.205255036Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=8d22d4d9-c241-4b10-a75e-d7c09c3d4ecb name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.217065620Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f" id=591e475f-8bc8-40da-bcc1-3895d96955f5 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.217296311Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:05fec0374aa998b4bcf349fcd2e1842b6b81578f4f5531c67be1ca535e80049a,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:16e6b7e6a929f4c75619c08ea9818f9be6c943a567e7739fee716da34d04950f],Size_:344931187,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=591e475f-8bc8-40da-bcc1-3895d96955f5 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.218500006Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-1/config-reloader" id=562b93d2-cf4a-47ac-9033-6b821ca21440 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.218612957Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.229123079Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=97190b4e-7971-4641-bb82-a6e0c6909ade name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.243435714Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=97190b4e-7971-4641-bb82-a6e0c6909ade name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.254966615Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=50e14291-d5a4-4d91-9edd-98aa4b7f0b0b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.255160965Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=50e14291-d5a4-4d91-9edd-98aa4b7f0b0b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.256000921Z" level=info msg="Creating container: openshift-monitoring/alertmanager-main-1/kube-rbac-proxy" id=56abda6d-3610-46f3-99d3-1c717e295b73 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.256115780Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.469931256Z" level=info msg="Created container 232339e500bf8d0b623fb2d890038c426b543e0aa41c526c4b18a11d5ab36d05: openshift-monitoring/prometheus-k8s-1/config-reloader" id=562b93d2-cf4a-47ac-9033-6b821ca21440 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.470451667Z" level=info msg="Starting container: 232339e500bf8d0b623fb2d890038c426b543e0aa41c526c4b18a11d5ab36d05" id=27e46a85-9428-446d-9cec-abc2b45c9c64 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.470670826Z" level=info msg="Created container ccff1752af95e2bb3a88430ecb83e316ddcd8bf48f73f5d938a17303185d0ef4: openshift-monitoring/alertmanager-main-1/kube-rbac-proxy" id=56abda6d-3610-46f3-99d3-1c717e295b73 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.473250489Z" level=info msg="Starting container: ccff1752af95e2bb3a88430ecb83e316ddcd8bf48f73f5d938a17303185d0ef4" id=60d81ed0-fbea-4355-a589-a5bbd318badb name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.497338912Z" level=info msg="Started container" PID=5846 containerID=232339e500bf8d0b623fb2d890038c426b543e0aa41c526c4b18a11d5ab36d05 description=openshift-monitoring/prometheus-k8s-1/config-reloader id=27e46a85-9428-446d-9cec-abc2b45c9c64 name=/runtime.v1.RuntimeService/StartContainer sandboxID=1df490500596677201ef8a2dc1ead29190963e5f81f951619c88d10237271737 Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.519740274Z" level=info msg="Started container" PID=5852 containerID=ccff1752af95e2bb3a88430ecb83e316ddcd8bf48f73f5d938a17303185d0ef4 description=openshift-monitoring/alertmanager-main-1/kube-rbac-proxy id=60d81ed0-fbea-4355-a589-a5bbd318badb name=/runtime.v1.RuntimeService/StartContainer sandboxID=6f2169b19a28ba86c3b28c7e7891c0a5cc09b988030c31c54bd83e2c69110f12 Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.528777702Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc" id=566c58f4-01d6-4fbd-9dab-33ccac1212cd name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.529071275Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:08ff7a43d9185529520a29e114f0c28b070000b464a44b347240015f3b466488,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc],Size_:368278089,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=566c58f4-01d6-4fbd-9dab-33ccac1212cd name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.532609247Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc" id=4c7adb07-6289-469c-a35c-acf32db4ad37 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.532813076Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:08ff7a43d9185529520a29e114f0c28b070000b464a44b347240015f3b466488,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:7f373268da226155f7be5e209ebeee7da910207c9870ad472246793521d73acc],Size_:368278089,Uid:nil,Username:nobody,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=4c7adb07-6289-469c-a35c-acf32db4ad37 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.533872221Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-1/thanos-sidecar" id=90810f8c-2558-44ec-8965-6ab3bc3c8b4f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.534010848Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.545980523Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=15991302-d0e4-4ef4-9a8b-f4b67b322f10 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.592113531Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=15991302-d0e4-4ef4-9a8b-f4b67b322f10 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.593121151Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=b5488152-e703-4a2a-9c93-9f4e2825aedc name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.593311219Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=b5488152-e703-4a2a-9c93-9f4e2825aedc name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.594433042Z" level=info msg="Creating container: openshift-monitoring/alertmanager-main-1/kube-rbac-proxy-metric" id=aa3ad637-d6df-4b31-a28d-979f168fd609 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:11.594596999Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:11.714470 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-1" event=&{ID:d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46 Type:ContainerStarted Data:ccff1752af95e2bb3a88430ecb83e316ddcd8bf48f73f5d938a17303185d0ef4} Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:11.714509 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-1" event=&{ID:d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46 Type:ContainerStarted Data:d3708188be0373e12e4347439a1a6bf7c8fa134f158a90035cbe341294b21505} Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:11.714525 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-1" event=&{ID:d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46 Type:ContainerStarted Data:68d3d2638852756ec6ba856e2401d3f254fdd1ac1a0e149de03ac08298f12bc1} Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:11.740430 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-1" event=&{ID:e5650cc7-8fac-496c-8567-d78340f40f5e Type:ContainerStarted Data:232339e500bf8d0b623fb2d890038c426b543e0aa41c526c4b18a11d5ab36d05} Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:11.740466 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-1" event=&{ID:e5650cc7-8fac-496c-8567-d78340f40f5e Type:ContainerStarted Data:9770e5949056ad84e03a67e5dfc68ea28b04d2c3058d8f91d131ef605477df7b} Jan 05 09:47:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:11.786498 1977 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/thanos-querier-6f4f5fb7cf-xg9rc" Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.059822555Z" level=info msg="Created container a3259a35746d136e2b057aeb4586469ee2875daaddf6b04e2a4b665306771a1a: openshift-monitoring/prometheus-k8s-1/thanos-sidecar" id=90810f8c-2558-44ec-8965-6ab3bc3c8b4f name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.060449717Z" level=info msg="Starting container: a3259a35746d136e2b057aeb4586469ee2875daaddf6b04e2a4b665306771a1a" id=0efa1078-3563-48af-8c51-98134d297ddb name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.094558930Z" level=info msg="Started container" PID=5925 containerID=a3259a35746d136e2b057aeb4586469ee2875daaddf6b04e2a4b665306771a1a description=openshift-monitoring/prometheus-k8s-1/thanos-sidecar id=0efa1078-3563-48af-8c51-98134d297ddb name=/runtime.v1.RuntimeService/StartContainer sandboxID=1df490500596677201ef8a2dc1ead29190963e5f81f951619c88d10237271737 Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.129033247Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=895f65c4-4f22-42dc-ae26-89700fc5181c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.129251334Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=895f65c4-4f22-42dc-ae26-89700fc5181c name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.130389457Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b" id=80987ddb-968f-4008-a158-731c45d0886b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.130576352Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:1882fc653fdd5021fb28f92ee94d40ce63a4bb370879d181ac973344d7b62a9e,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b6536bfcfaf30a6425d589facd672bae3245f933b2a7399bda3f12e393bd671b],Size_:377361805,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=80987ddb-968f-4008-a158-731c45d0886b name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.131860905Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-1/prometheus-proxy" id=c5e6e03d-db39-4712-9c81-f479f3d56d79 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.131994491Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.319694482Z" level=info msg="Created container 7a6e5e7e17130176958893707090b3a805445ac8a504d33c33ab0c04bdc5b3b8: openshift-monitoring/alertmanager-main-1/kube-rbac-proxy-metric" id=aa3ad637-d6df-4b31-a28d-979f168fd609 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.331397431Z" level=info msg="Starting container: 7a6e5e7e17130176958893707090b3a805445ac8a504d33c33ab0c04bdc5b3b8" id=4b8e90ab-061a-496c-b8d0-a4cbb2bb89f8 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.362917618Z" level=info msg="Started container" PID=5939 containerID=7a6e5e7e17130176958893707090b3a805445ac8a504d33c33ab0c04bdc5b3b8 description=openshift-monitoring/alertmanager-main-1/kube-rbac-proxy-metric id=4b8e90ab-061a-496c-b8d0-a4cbb2bb89f8 name=/runtime.v1.RuntimeService/StartContainer sandboxID=6f2169b19a28ba86c3b28c7e7891c0a5cc09b988030c31c54bd83e2c69110f12 Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.412541036Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816" id=75d1d914-e5a3-4a33-bb09-a3ecd1342dac name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.412783236Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:955ee2b41ecd88e3a9c0183a0b65d6edbf524cefa1318837dcbc82b0466c4d55,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816],Size_:309576938,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=75d1d914-e5a3-4a33-bb09-a3ecd1342dac name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.413724574Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816" id=96c033e9-5756-4361-941e-9fb12824301a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.413943080Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:955ee2b41ecd88e3a9c0183a0b65d6edbf524cefa1318837dcbc82b0466c4d55,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:63a03be7e8759270b4a2c75112207867ef51774db5f682d4d2f711d0934fc816],Size_:309576938,Uid:&Int64Value{Value:1001,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=96c033e9-5756-4361-941e-9fb12824301a name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.414791615Z" level=info msg="Creating container: openshift-monitoring/alertmanager-main-1/prom-label-proxy" id=4bcc5ff5-7541-4513-9562-3926cb963b9a name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.414923519Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.617453175Z" level=info msg="Created container 3c9b28f523fd918e20a7c7fa68219008087ce6189a730a50cf4d5d12cb2ef623: openshift-monitoring/prometheus-k8s-1/prometheus-proxy" id=c5e6e03d-db39-4712-9c81-f479f3d56d79 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.617987094Z" level=info msg="Starting container: 3c9b28f523fd918e20a7c7fa68219008087ce6189a730a50cf4d5d12cb2ef623" id=c6821d98-e945-419b-ab5d-1d86bbe7214b name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.659365920Z" level=info msg="Started container" PID=6017 containerID=3c9b28f523fd918e20a7c7fa68219008087ce6189a730a50cf4d5d12cb2ef623 description=openshift-monitoring/prometheus-k8s-1/prometheus-proxy id=c6821d98-e945-419b-ab5d-1d86bbe7214b name=/runtime.v1.RuntimeService/StartContainer sandboxID=1df490500596677201ef8a2dc1ead29190963e5f81f951619c88d10237271737 Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.704379041Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=ffb215d5-b298-49f4-bce5-0fbc3fb79aee name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.704615869Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=ffb215d5-b298-49f4-bce5-0fbc3fb79aee name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.705681698Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=5c452304-0b39-479d-913b-59e095c40441 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.705892055Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=5c452304-0b39-479d-913b-59e095c40441 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.707188709Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-1/kube-rbac-proxy" id=2f4b7b5c-2cbc-486f-ae32-bff288911cd0 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.707331369Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:12.744050 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-1" event=&{ID:d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46 Type:ContainerStarted Data:7a6e5e7e17130176958893707090b3a805445ac8a504d33c33ab0c04bdc5b3b8} Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:12.748518 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-1" event=&{ID:e5650cc7-8fac-496c-8567-d78340f40f5e Type:ContainerStarted Data:3c9b28f523fd918e20a7c7fa68219008087ce6189a730a50cf4d5d12cb2ef623} Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:12.748549 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-1" event=&{ID:e5650cc7-8fac-496c-8567-d78340f40f5e Type:ContainerStarted Data:a3259a35746d136e2b057aeb4586469ee2875daaddf6b04e2a4b665306771a1a} Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.881087463Z" level=info msg="Created container 13a3a591a30e5ce31e7a8f1dc4a455f81ca837c3e942853154c12c2776d63f22: openshift-monitoring/alertmanager-main-1/prom-label-proxy" id=4bcc5ff5-7541-4513-9562-3926cb963b9a name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.881675474Z" level=info msg="Starting container: 13a3a591a30e5ce31e7a8f1dc4a455f81ca837c3e942853154c12c2776d63f22" id=51952225-2ce1-42e8-88ed-70d9bc9ef2f7 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.908024780Z" level=info msg="Started container" PID=6061 containerID=13a3a591a30e5ce31e7a8f1dc4a455f81ca837c3e942853154c12c2776d63f22 description=openshift-monitoring/alertmanager-main-1/prom-label-proxy id=51952225-2ce1-42e8-88ed-70d9bc9ef2f7 name=/runtime.v1.RuntimeService/StartContainer sandboxID=6f2169b19a28ba86c3b28c7e7891c0a5cc09b988030c31c54bd83e2c69110f12 Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.945397019Z" level=info msg="Created container 55ebb8ae543110a1fdbd62b133c9180a875a8431a8362d855df08f7c9ac46ea8: openshift-monitoring/prometheus-k8s-1/kube-rbac-proxy" id=2f4b7b5c-2cbc-486f-ae32-bff288911cd0 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.945782415Z" level=info msg="Starting container: 55ebb8ae543110a1fdbd62b133c9180a875a8431a8362d855df08f7c9ac46ea8" id=e15db344-6153-4da9-910f-7b161601fa68 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.952737215Z" level=info msg="Started container" PID=6089 containerID=55ebb8ae543110a1fdbd62b133c9180a875a8431a8362d855df08f7c9ac46ea8 description=openshift-monitoring/prometheus-k8s-1/kube-rbac-proxy id=e15db344-6153-4da9-910f-7b161601fa68 name=/runtime.v1.RuntimeService/StartContainer sandboxID=1df490500596677201ef8a2dc1ead29190963e5f81f951619c88d10237271737 Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.963996255Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=bb964f08-25f1-4b4f-86e1-0cac38a1186d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.964218557Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=bb964f08-25f1-4b4f-86e1-0cac38a1186d name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.965577399Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9" id=911a5f18-377c-4c3c-b09c-407ee8b2d11f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.965770449Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:3d2469faedce2d14a507014382f19da586e6288da5e4eb9ee4376ea5999a1627,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:ff07afaff7fff38e3124162c5791ba969bcd750be02d5878f0387aed62517fa9],Size_:433414846,Uid:&Int64Value{Value:65534,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=911a5f18-377c-4c3c-b09c-407ee8b2d11f name=/runtime.v1.ImageService/ImageStatus Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.966852907Z" level=info msg="Creating container: openshift-monitoring/prometheus-k8s-1/kube-rbac-proxy-thanos" id=aa528735-8cd3-463b-97ce-be5c9995cbe0 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:12.966986643Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:47:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:13.082648878Z" level=info msg="Created container 9bed50d0f908716890af0e09369588317d2b65245f559d1a7989705cb34e6aed: openshift-monitoring/prometheus-k8s-1/kube-rbac-proxy-thanos" id=aa528735-8cd3-463b-97ce-be5c9995cbe0 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:47:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:13.083120206Z" level=info msg="Starting container: 9bed50d0f908716890af0e09369588317d2b65245f559d1a7989705cb34e6aed" id=35076bfe-02db-46c1-9bce-b8895c7d8c76 name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:47:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:13.102889293Z" level=info msg="Started container" PID=6143 containerID=9bed50d0f908716890af0e09369588317d2b65245f559d1a7989705cb34e6aed description=openshift-monitoring/prometheus-k8s-1/kube-rbac-proxy-thanos id=35076bfe-02db-46c1-9bce-b8895c7d8c76 name=/runtime.v1.RuntimeService/StartContainer sandboxID=1df490500596677201ef8a2dc1ead29190963e5f81f951619c88d10237271737 Jan 05 09:47:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:13.750637 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-1" event=&{ID:e5650cc7-8fac-496c-8567-d78340f40f5e Type:ContainerStarted Data:9bed50d0f908716890af0e09369588317d2b65245f559d1a7989705cb34e6aed} Jan 05 09:47:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:13.750671 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-1" event=&{ID:e5650cc7-8fac-496c-8567-d78340f40f5e Type:ContainerStarted Data:55ebb8ae543110a1fdbd62b133c9180a875a8431a8362d855df08f7c9ac46ea8} Jan 05 09:47:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:13.753189 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-1" event=&{ID:d37fa6e9-f7ae-46f4-bf1b-d6e57f095d46 Type:ContainerStarted Data:13a3a591a30e5ce31e7a8f1dc4a455f81ca837c3e942853154c12c2776d63f22} Jan 05 09:47:15 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:15.012401 1977 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xmq2g" Jan 05 09:47:17 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:17.811349 1977 kubelet.go:2229] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:18 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:18.312484 1977 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:23.312646 1977 kubelet.go:2229] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:23.359369 1977 kubelet.go:2229] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:23 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:23.815304 1977 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-k8s-1" Jan 05 09:47:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:47:27.633117581Z" level=warning msg="Found defunct process with PID 6224 (haproxy)" Jan 05 09:47:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:37.850406 1977 kubelet.go:2229] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/alertmanager-main-1" Jan 05 09:47:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:47:40.013285 1977 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-adapter-5894cc86c7-rxkb2" Jan 05 09:49:06 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:49:06.487962215Z" level=warning msg="Found defunct process with PID 7872 (haproxy)" Jan 05 09:49:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:49:27.634285807Z" level=warning msg="Found defunct process with PID 9171 (haproxy)" Jan 05 09:49:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:49:36.075048 1977 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-multus/cni-sysctl-allowlist-ds-svrcn] Jan 05 09:49:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:49:36.075100 1977 topology_manager.go:205] "Topology Admit Handler" Jan 05 09:49:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:49:36.231931 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/31e7d900-94e8-4153-9a9d-c747e858429d-ready\") pod \"cni-sysctl-allowlist-ds-svrcn\" (UID: \"31e7d900-94e8-4153-9a9d-c747e858429d\") " pod="openshift-multus/cni-sysctl-allowlist-ds-svrcn" Jan 05 09:49:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:49:36.231995 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/31e7d900-94e8-4153-9a9d-c747e858429d-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-svrcn\" (UID: \"31e7d900-94e8-4153-9a9d-c747e858429d\") " pod="openshift-multus/cni-sysctl-allowlist-ds-svrcn" Jan 05 09:49:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:49:36.232047 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65zzh\" (UniqueName: \"kubernetes.io/projected/31e7d900-94e8-4153-9a9d-c747e858429d-kube-api-access-65zzh\") pod \"cni-sysctl-allowlist-ds-svrcn\" (UID: \"31e7d900-94e8-4153-9a9d-c747e858429d\") " pod="openshift-multus/cni-sysctl-allowlist-ds-svrcn" Jan 05 09:49:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:49:36.232084 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/31e7d900-94e8-4153-9a9d-c747e858429d-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-svrcn\" (UID: \"31e7d900-94e8-4153-9a9d-c747e858429d\") " pod="openshift-multus/cni-sysctl-allowlist-ds-svrcn" Jan 05 09:49:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:49:36.332788 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/31e7d900-94e8-4153-9a9d-c747e858429d-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-svrcn\" (UID: \"31e7d900-94e8-4153-9a9d-c747e858429d\") " pod="openshift-multus/cni-sysctl-allowlist-ds-svrcn" Jan 05 09:49:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:49:36.332882 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-65zzh\" (UniqueName: \"kubernetes.io/projected/31e7d900-94e8-4153-9a9d-c747e858429d-kube-api-access-65zzh\") pod \"cni-sysctl-allowlist-ds-svrcn\" (UID: \"31e7d900-94e8-4153-9a9d-c747e858429d\") " pod="openshift-multus/cni-sysctl-allowlist-ds-svrcn" Jan 05 09:49:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:49:36.332951 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/31e7d900-94e8-4153-9a9d-c747e858429d-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-svrcn\" (UID: \"31e7d900-94e8-4153-9a9d-c747e858429d\") " pod="openshift-multus/cni-sysctl-allowlist-ds-svrcn" Jan 05 09:49:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:49:36.332986 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/31e7d900-94e8-4153-9a9d-c747e858429d-ready\") pod \"cni-sysctl-allowlist-ds-svrcn\" (UID: \"31e7d900-94e8-4153-9a9d-c747e858429d\") " pod="openshift-multus/cni-sysctl-allowlist-ds-svrcn" Jan 05 09:49:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:49:36.333338 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/31e7d900-94e8-4153-9a9d-c747e858429d-ready\") pod \"cni-sysctl-allowlist-ds-svrcn\" (UID: \"31e7d900-94e8-4153-9a9d-c747e858429d\") " pod="openshift-multus/cni-sysctl-allowlist-ds-svrcn" Jan 05 09:49:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:49:36.333742 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/31e7d900-94e8-4153-9a9d-c747e858429d-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-svrcn\" (UID: \"31e7d900-94e8-4153-9a9d-c747e858429d\") " pod="openshift-multus/cni-sysctl-allowlist-ds-svrcn" Jan 05 09:49:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:49:36.334081 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/31e7d900-94e8-4153-9a9d-c747e858429d-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-svrcn\" (UID: \"31e7d900-94e8-4153-9a9d-c747e858429d\") " pod="openshift-multus/cni-sysctl-allowlist-ds-svrcn" Jan 05 09:49:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:49:36.351341 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-65zzh\" (UniqueName: \"kubernetes.io/projected/31e7d900-94e8-4153-9a9d-c747e858429d-kube-api-access-65zzh\") pod \"cni-sysctl-allowlist-ds-svrcn\" (UID: \"31e7d900-94e8-4153-9a9d-c747e858429d\") " pod="openshift-multus/cni-sysctl-allowlist-ds-svrcn" Jan 05 09:49:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:49:36.407252 1977 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-svrcn" Jan 05 09:49:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:49:36.407668425Z" level=info msg="Running pod sandbox: openshift-multus/cni-sysctl-allowlist-ds-svrcn/POD" id=28901814-a586-4d6b-b074-a3c2eaa8d880 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:49:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:49:36.407722995Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:49:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:49:36.429278589Z" level=info msg="Got pod network &{Name:cni-sysctl-allowlist-ds-svrcn Namespace:openshift-multus ID:238a62e61937b1d17028fc0a1c1468dfdc53d037a8634998623d2fca3078712e UID:31e7d900-94e8-4153-9a9d-c747e858429d NetNS:/var/run/netns/481d2f43-7642-446d-8e3f-e8e5240a3953 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:49:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:49:36.429313337Z" level=info msg="Adding pod openshift-multus_cni-sysctl-allowlist-ds-svrcn to CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:49:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:49:36.491359043Z" level=warning msg="Found defunct process with PID 9628 (haproxy)" Jan 05 09:49:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:49:36.491603573Z" level=warning msg="Found defunct process with PID 9868 (systemd-sysctl)" Jan 05 09:49:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:49:37.713521 1977 kubelet.go:2126] "SyncLoop UPDATE" source="api" pods=[openshift-multus/cni-sysctl-allowlist-ds-svrcn] Jan 05 09:49:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: 2023-01-05T09:49:37Z [verbose] Add: openshift-multus:cni-sysctl-allowlist-ds-svrcn:31e7d900-94e8-4153-9a9d-c747e858429d:openshift-sdn(openshift-sdn):eth0 {"cniVersion":"0.3.1","interfaces":[{"name":"eth0","sandbox":"/var/run/netns/481d2f43-7642-446d-8e3f-e8e5240a3953"}],"ips":[{"version":"4","interface":0,"address":"10.131.0.13/23","gateway":"10.131.0.1"}],"dns":{}} Jan 05 09:49:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: I0105 09:49:37.695221 9856 event.go:282] Event(v1.ObjectReference{Kind:"Pod", Namespace:"openshift-multus", Name:"cni-sysctl-allowlist-ds-svrcn", UID:"31e7d900-94e8-4153-9a9d-c747e858429d", APIVersion:"v1", ResourceVersion:"111026", FieldPath:""}): type: 'Normal' reason: 'AddedInterface' Add eth0 [10.131.0.13/23] from openshift-sdn Jan 05 09:49:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:49:37.715026153Z" level=info msg="Got pod network &{Name:cni-sysctl-allowlist-ds-svrcn Namespace:openshift-multus ID:238a62e61937b1d17028fc0a1c1468dfdc53d037a8634998623d2fca3078712e UID:31e7d900-94e8-4153-9a9d-c747e858429d NetNS:/var/run/netns/481d2f43-7642-446d-8e3f-e8e5240a3953 Networks:[] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:49:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:49:37.715170489Z" level=info msg="Checking pod openshift-multus_cni-sysctl-allowlist-ds-svrcn for CNI network multus-cni-network (type=multus)" Jan 05 09:49:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: W0105 09:49:37.723966 1977 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31e7d900_94e8_4153_9a9d_c747e858429d.slice/crio-238a62e61937b1d17028fc0a1c1468dfdc53d037a8634998623d2fca3078712e.scope WatchSource:0}: Error finding container 238a62e61937b1d17028fc0a1c1468dfdc53d037a8634998623d2fca3078712e: Status 404 returned error can't find the container with id 238a62e61937b1d17028fc0a1c1468dfdc53d037a8634998623d2fca3078712e Jan 05 09:49:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:49:37.728156331Z" level=info msg="Ran pod sandbox 238a62e61937b1d17028fc0a1c1468dfdc53d037a8634998623d2fca3078712e with infra container: openshift-multus/cni-sysctl-allowlist-ds-svrcn/POD" id=28901814-a586-4d6b-b074-a3c2eaa8d880 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 09:49:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:49:37.729129883Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=80277667-5c05-4ff8-bf30-8b12d2ccf8e5 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:49:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:49:37.729364330Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=80277667-5c05-4ff8-bf30-8b12d2ccf8e5 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:49:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:49:37.733090554Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f" id=156f811e-cf77-4caa-92f5-10ee95561cdc name=/runtime.v1.ImageService/ImageStatus Jan 05 09:49:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:49:37.733269734Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:f1434dedc3089136ed080b74deb0603ecd12ab59f26155bb84528bdd143196fe,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:cc6f52aff25ba6cad2df5f4dc4ca65abcb0e46880438397646d194b9e595001f],Size_:414527192,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=156f811e-cf77-4caa-92f5-10ee95561cdc name=/runtime.v1.ImageService/ImageStatus Jan 05 09:49:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:49:37.737735100Z" level=info msg="Creating container: openshift-multus/cni-sysctl-allowlist-ds-svrcn/kube-multus-additional-cni-plugins" id=c3c174b7-df46-423b-b078-7713346954ff name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:49:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:49:37.737890587Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 09:49:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:49:37.996525976Z" level=info msg="Created container 2029504da2e99fe6583dde7b2e1631adcc8a7c55d0e96f63eaeed27b265c1fb2: openshift-multus/cni-sysctl-allowlist-ds-svrcn/kube-multus-additional-cni-plugins" id=c3c174b7-df46-423b-b078-7713346954ff name=/runtime.v1.RuntimeService/CreateContainer Jan 05 09:49:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:49:37.997376206Z" level=info msg="Starting container: 2029504da2e99fe6583dde7b2e1631adcc8a7c55d0e96f63eaeed27b265c1fb2" id=a0ff55e4-a1f0-4e6c-898c-b6e7d5262fef name=/runtime.v1.RuntimeService/StartContainer Jan 05 09:49:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:49:38.016344257Z" level=info msg="Started container" PID=9940 containerID=2029504da2e99fe6583dde7b2e1631adcc8a7c55d0e96f63eaeed27b265c1fb2 description=openshift-multus/cni-sysctl-allowlist-ds-svrcn/kube-multus-additional-cni-plugins id=a0ff55e4-a1f0-4e6c-898c-b6e7d5262fef name=/runtime.v1.RuntimeService/StartContainer sandboxID=238a62e61937b1d17028fc0a1c1468dfdc53d037a8634998623d2fca3078712e Jan 05 09:49:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:49:38.136276 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-svrcn" event=&{ID:31e7d900-94e8-4153-9a9d-c747e858429d Type:ContainerStarted Data:2029504da2e99fe6583dde7b2e1631adcc8a7c55d0e96f63eaeed27b265c1fb2} Jan 05 09:49:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:49:38.136308 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-svrcn" event=&{ID:31e7d900-94e8-4153-9a9d-c747e858429d Type:ContainerStarted Data:238a62e61937b1d17028fc0a1c1468dfdc53d037a8634998623d2fca3078712e} Jan 05 09:49:38 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:49:38.136422 1977 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-multus/cni-sysctl-allowlist-ds-svrcn" Jan 05 09:49:39 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:49:39.172415 1977 kubelet.go:2229] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-multus/cni-sysctl-allowlist-ds-svrcn" Jan 05 09:49:40 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:49:40.664643 1977 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-multus/cni-sysctl-allowlist-ds-svrcn] Jan 05 09:49:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:49:41.142250 1977 kuberuntime_container.go:702] "Killing container with a grace period" pod="openshift-multus/cni-sysctl-allowlist-ds-svrcn" podUID=31e7d900-94e8-4153-9a9d-c747e858429d containerName="kube-multus-additional-cni-plugins" containerID="cri-o://2029504da2e99fe6583dde7b2e1631adcc8a7c55d0e96f63eaeed27b265c1fb2" gracePeriod=30 Jan 05 09:49:41 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:49:41.142655257Z" level=info msg="Stopping container: 2029504da2e99fe6583dde7b2e1631adcc8a7c55d0e96f63eaeed27b265c1fb2 (timeout: 30s)" id=e2acbb11-b6fd-4954-a575-192ba7f6c472 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:50:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:50:11.150697156Z" level=warning msg="Stopping container 2029504da2e99fe6583dde7b2e1631adcc8a7c55d0e96f63eaeed27b265c1fb2 with stop signal timed out: timeout reached after 30 seconds waiting for container process to exit" id=e2acbb11-b6fd-4954-a575-192ba7f6c472 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:50:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:50:11.297262801Z" level=info msg="Stopped container 2029504da2e99fe6583dde7b2e1631adcc8a7c55d0e96f63eaeed27b265c1fb2: openshift-multus/cni-sysctl-allowlist-ds-svrcn/kube-multus-additional-cni-plugins" id=e2acbb11-b6fd-4954-a575-192ba7f6c472 name=/runtime.v1.RuntimeService/StopContainer Jan 05 09:50:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:50:11.297695702Z" level=info msg="Stopping pod sandbox: 238a62e61937b1d17028fc0a1c1468dfdc53d037a8634998623d2fca3078712e" id=72575040-1602-43bc-8142-c0947c502748 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:50:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:50:11.298040889Z" level=info msg="Got pod network &{Name:cni-sysctl-allowlist-ds-svrcn Namespace:openshift-multus ID:238a62e61937b1d17028fc0a1c1468dfdc53d037a8634998623d2fca3078712e UID:31e7d900-94e8-4153-9a9d-c747e858429d NetNS:/var/run/netns/481d2f43-7642-446d-8e3f-e8e5240a3953 Networks:[{Name:multus-cni-network Ifname:eth0}] RuntimeConfig:map[multus-cni-network:{IP: MAC: PortMappings:[] Bandwidth: IpRanges:[]}] Aliases:map[]}" Jan 05 09:50:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:50:11.298167801Z" level=info msg="Deleting pod openshift-multus_cni-sysctl-allowlist-ds-svrcn from CNI network \"multus-cni-network\" (type=multus)" Jan 05 09:50:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: 2023-01-05T09:50:11Z [verbose] Del: openshift-multus:cni-sysctl-allowlist-ds-svrcn:31e7d900-94e8-4153-9a9d-c747e858429d:openshift-sdn:eth0 {"cniVersion":"0.3.1","name":"openshift-sdn","type":"openshift-sdn"} Jan 05 09:50:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:50:11.462006170Z" level=info msg="Stopped pod sandbox: 238a62e61937b1d17028fc0a1c1468dfdc53d037a8634998623d2fca3078712e" id=72575040-1602-43bc-8142-c0947c502748 name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:50:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:50:11.664545 1977 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65zzh\" (UniqueName: \"kubernetes.io/projected/31e7d900-94e8-4153-9a9d-c747e858429d-kube-api-access-65zzh\") pod \"31e7d900-94e8-4153-9a9d-c747e858429d\" (UID: \"31e7d900-94e8-4153-9a9d-c747e858429d\") " Jan 05 09:50:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:50:11.664598 1977 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/31e7d900-94e8-4153-9a9d-c747e858429d-ready\") pod \"31e7d900-94e8-4153-9a9d-c747e858429d\" (UID: \"31e7d900-94e8-4153-9a9d-c747e858429d\") " Jan 05 09:50:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:50:11.664630 1977 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/31e7d900-94e8-4153-9a9d-c747e858429d-tuning-conf-dir\") pod \"31e7d900-94e8-4153-9a9d-c747e858429d\" (UID: \"31e7d900-94e8-4153-9a9d-c747e858429d\") " Jan 05 09:50:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:50:11.664686 1977 reconciler.go:211] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/31e7d900-94e8-4153-9a9d-c747e858429d-cni-sysctl-allowlist\") pod \"31e7d900-94e8-4153-9a9d-c747e858429d\" (UID: \"31e7d900-94e8-4153-9a9d-c747e858429d\") " Jan 05 09:50:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: W0105 09:50:11.665030 1977 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/31e7d900-94e8-4153-9a9d-c747e858429d/volumes/kubernetes.io~configmap/cni-sysctl-allowlist: clearQuota called, but quotas disabled Jan 05 09:50:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:50:11.665247 1977 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31e7d900-94e8-4153-9a9d-c747e858429d-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "31e7d900-94e8-4153-9a9d-c747e858429d" (UID: "31e7d900-94e8-4153-9a9d-c747e858429d"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 05 09:50:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:50:11.665296 1977 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/31e7d900-94e8-4153-9a9d-c747e858429d-tuning-conf-dir" (OuterVolumeSpecName: "tuning-conf-dir") pod "31e7d900-94e8-4153-9a9d-c747e858429d" (UID: "31e7d900-94e8-4153-9a9d-c747e858429d"). InnerVolumeSpecName "tuning-conf-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 05 09:50:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: W0105 09:50:11.665315 1977 empty_dir.go:523] Warning: Failed to clear quota on /var/lib/kubelet/pods/31e7d900-94e8-4153-9a9d-c747e858429d/volumes/kubernetes.io~empty-dir/ready: clearQuota called, but quotas disabled Jan 05 09:50:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:50:11.665398 1977 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31e7d900-94e8-4153-9a9d-c747e858429d-ready" (OuterVolumeSpecName: "ready") pod "31e7d900-94e8-4153-9a9d-c747e858429d" (UID: "31e7d900-94e8-4153-9a9d-c747e858429d"). InnerVolumeSpecName "ready". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 05 09:50:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:50:11.672197 1977 operation_generator.go:890] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31e7d900-94e8-4153-9a9d-c747e858429d-kube-api-access-65zzh" (OuterVolumeSpecName: "kube-api-access-65zzh") pod "31e7d900-94e8-4153-9a9d-c747e858429d" (UID: "31e7d900-94e8-4153-9a9d-c747e858429d"). InnerVolumeSpecName "kube-api-access-65zzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 05 09:50:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:50:11.765611 1977 reconciler.go:399] "Volume detached for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/31e7d900-94e8-4153-9a9d-c747e858429d-ready\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:50:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:50:11.765648 1977 reconciler.go:399] "Volume detached for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/31e7d900-94e8-4153-9a9d-c747e858429d-tuning-conf-dir\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:50:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:50:11.765666 1977 reconciler.go:399] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/31e7d900-94e8-4153-9a9d-c747e858429d-cni-sysctl-allowlist\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:50:11 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:50:11.765681 1977 reconciler.go:399] "Volume detached for volume \"kube-api-access-65zzh\" (UniqueName: \"kubernetes.io/projected/31e7d900-94e8-4153-9a9d-c747e858429d-kube-api-access-65zzh\") on node \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl\" DevicePath \"\"" Jan 05 09:50:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:50:12.249965 1977 generic.go:296] "Generic (PLEG): container finished" podID=31e7d900-94e8-4153-9a9d-c747e858429d containerID="2029504da2e99fe6583dde7b2e1631adcc8a7c55d0e96f63eaeed27b265c1fb2" exitCode=137 Jan 05 09:50:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:50:12.250001 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-svrcn" event=&{ID:31e7d900-94e8-4153-9a9d-c747e858429d Type:ContainerDied Data:2029504da2e99fe6583dde7b2e1631adcc8a7c55d0e96f63eaeed27b265c1fb2} Jan 05 09:50:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:50:12.250026 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-svrcn" event=&{ID:31e7d900-94e8-4153-9a9d-c747e858429d Type:ContainerDied Data:238a62e61937b1d17028fc0a1c1468dfdc53d037a8634998623d2fca3078712e} Jan 05 09:50:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:50:12.250060 1977 scope.go:115] "RemoveContainer" containerID="2029504da2e99fe6583dde7b2e1631adcc8a7c55d0e96f63eaeed27b265c1fb2" Jan 05 09:50:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:50:12.255792085Z" level=info msg="Removing container: 2029504da2e99fe6583dde7b2e1631adcc8a7c55d0e96f63eaeed27b265c1fb2" id=c9116d72-2d08-44cf-9ecf-820410daf35b name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:50:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:50:12.281084268Z" level=info msg="Removed container 2029504da2e99fe6583dde7b2e1631adcc8a7c55d0e96f63eaeed27b265c1fb2: openshift-multus/cni-sysctl-allowlist-ds-svrcn/kube-multus-additional-cni-plugins" id=c9116d72-2d08-44cf-9ecf-820410daf35b name=/runtime.v1.RuntimeService/RemoveContainer Jan 05 09:50:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:50:12.281371 1977 scope.go:115] "RemoveContainer" containerID="2029504da2e99fe6583dde7b2e1631adcc8a7c55d0e96f63eaeed27b265c1fb2" Jan 05 09:50:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: E0105 09:50:12.281638 1977 remote_runtime.go:625] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2029504da2e99fe6583dde7b2e1631adcc8a7c55d0e96f63eaeed27b265c1fb2\": container with ID starting with 2029504da2e99fe6583dde7b2e1631adcc8a7c55d0e96f63eaeed27b265c1fb2 not found: ID does not exist" containerID="2029504da2e99fe6583dde7b2e1631adcc8a7c55d0e96f63eaeed27b265c1fb2" Jan 05 09:50:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:50:12.281680 1977 pod_container_deletor.go:52] "DeleteContainer returned error" containerID={Type:cri-o ID:2029504da2e99fe6583dde7b2e1631adcc8a7c55d0e96f63eaeed27b265c1fb2} err="failed to get container status \"2029504da2e99fe6583dde7b2e1631adcc8a7c55d0e96f63eaeed27b265c1fb2\": rpc error: code = NotFound desc = could not find container \"2029504da2e99fe6583dde7b2e1631adcc8a7c55d0e96f63eaeed27b265c1fb2\": container with ID starting with 2029504da2e99fe6583dde7b2e1631adcc8a7c55d0e96f63eaeed27b265c1fb2 not found: ID does not exist" Jan 05 09:50:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:50:12.285801 1977 kubelet.go:2135] "SyncLoop DELETE" source="api" pods=[openshift-multus/cni-sysctl-allowlist-ds-svrcn] Jan 05 09:50:12 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:50:12.287758 1977 kubelet.go:2129] "SyncLoop REMOVE" source="api" pods=[openshift-multus/cni-sysctl-allowlist-ds-svrcn] Jan 05 09:50:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:50:13.286263640Z" level=info msg="Stopping pod sandbox: 238a62e61937b1d17028fc0a1c1468dfdc53d037a8634998623d2fca3078712e" id=81f61fc6-a2d9-4180-938a-265f6ceb3b6d name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:50:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:50:13.286304285Z" level=info msg="Stopped pod sandbox (already stopped): 238a62e61937b1d17028fc0a1c1468dfdc53d037a8634998623d2fca3078712e" id=81f61fc6-a2d9-4180-938a-265f6ceb3b6d name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:50:13 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:50:13.288599 1977 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=31e7d900-94e8-4153-9a9d-c747e858429d path="/var/lib/kubelet/pods/31e7d900-94e8-4153-9a9d-c747e858429d/volumes" Jan 05 09:50:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:50:37.028607364Z" level=info msg="Stopping pod sandbox: 238a62e61937b1d17028fc0a1c1468dfdc53d037a8634998623d2fca3078712e" id=7628880c-6ad3-4e70-97d8-66b6bb6c9c8a name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:50:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:50:37.028653561Z" level=info msg="Stopped pod sandbox (already stopped): 238a62e61937b1d17028fc0a1c1468dfdc53d037a8634998623d2fca3078712e" id=7628880c-6ad3-4e70-97d8-66b6bb6c9c8a name=/runtime.v1.RuntimeService/StopPodSandbox Jan 05 09:50:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:50:37.029018976Z" level=info msg="Removing pod sandbox: 238a62e61937b1d17028fc0a1c1468dfdc53d037a8634998623d2fca3078712e" id=6bdef0eb-64ee-4aa9-bbb4-171fd4b57d57 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:50:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:50:37.036471921Z" level=info msg="Removed pod sandbox: 238a62e61937b1d17028fc0a1c1468dfdc53d037a8634998623d2fca3078712e" id=6bdef0eb-64ee-4aa9-bbb4-171fd4b57d57 name=/runtime.v1.RuntimeService/RemovePodSandbox Jan 05 09:51:36 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 09:51:36.933871 1977 kubelet.go:1343] "Image garbage collection succeeded" Jan 05 09:51:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:51:37.201000252Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=2d677140-77c1-4324-86dd-e47f14ca0164 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:51:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:51:37.201259894Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=2d677140-77c1-4324-86dd-e47f14ca0164 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:52:27 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:52:27.632856962Z" level=warning msg="Found defunct process with PID 9404 (haproxy)" Jan 05 09:56:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:56:37.204915063Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=0a40a9a3-0b8f-49f1-bf07-ce3faef3a675 name=/runtime.v1.ImageService/ImageStatus Jan 05 09:56:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 09:56:37.205155595Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=0a40a9a3-0b8f-49f1-bf07-ce3faef3a675 name=/runtime.v1.ImageService/ImageStatus Jan 05 10:01:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 10:01:37.208576421Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12" id=73028239-71ba-42ca-93be-6409285b332e name=/runtime.v1.ImageService/ImageStatus Jan 05 10:01:37 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 10:01:37.208893274Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:2e45f897d712b887979685fa5a3c6eeffd8a870592a8a5b18a740385970729ad,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:783217430f1b5c4c56cfc28d5bd2a58ec36ddc200e091bc97dda1a1150db7b12],Size_:351746909,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=73028239-71ba-42ca-93be-6409285b332e name=/runtime.v1.ImageService/ImageStatus Jan 05 10:02:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 10:02:50.344586 1977 kubelet.go:2119] "SyncLoop ADD" source="api" pods=[openshift-debug-4sl8v/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl-debug] Jan 05 10:02:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 10:02:50.344638 1977 topology_manager.go:205] "Topology Admit Handler" Jan 05 10:02:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: E0105 10:02:50.344743 1977 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="31e7d900-94e8-4153-9a9d-c747e858429d" containerName="kube-multus-additional-cni-plugins" Jan 05 10:02:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 10:02:50.344755 1977 state_mem.go:107] "Deleted CPUSet assignment" podUID="31e7d900-94e8-4153-9a9d-c747e858429d" containerName="kube-multus-additional-cni-plugins" Jan 05 10:02:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 10:02:50.344823 1977 memory_manager.go:345] "RemoveStaleState removing state" podUID="31e7d900-94e8-4153-9a9d-c747e858429d" containerName="kube-multus-additional-cni-plugins" Jan 05 10:02:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 10:02:50.455763 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcwnr\" (UniqueName: \"kubernetes.io/projected/281d0fef-ebf1-4d7c-a906-f91e1432adbf-kube-api-access-gcwnr\") pod \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl-debug\" (UID: \"281d0fef-ebf1-4d7c-a906-f91e1432adbf\") " pod="openshift-debug-4sl8v/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl-debug" Jan 05 10:02:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 10:02:50.455828 1977 reconciler.go:357] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/281d0fef-ebf1-4d7c-a906-f91e1432adbf-host\") pod \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl-debug\" (UID: \"281d0fef-ebf1-4d7c-a906-f91e1432adbf\") " pod="openshift-debug-4sl8v/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl-debug" Jan 05 10:02:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 10:02:50.557054 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"kube-api-access-gcwnr\" (UniqueName: \"kubernetes.io/projected/281d0fef-ebf1-4d7c-a906-f91e1432adbf-kube-api-access-gcwnr\") pod \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl-debug\" (UID: \"281d0fef-ebf1-4d7c-a906-f91e1432adbf\") " pod="openshift-debug-4sl8v/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl-debug" Jan 05 10:02:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 10:02:50.557117 1977 reconciler.go:269] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/281d0fef-ebf1-4d7c-a906-f91e1432adbf-host\") pod \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl-debug\" (UID: \"281d0fef-ebf1-4d7c-a906-f91e1432adbf\") " pod="openshift-debug-4sl8v/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl-debug" Jan 05 10:02:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 10:02:50.557230 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/281d0fef-ebf1-4d7c-a906-f91e1432adbf-host\") pod \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl-debug\" (UID: \"281d0fef-ebf1-4d7c-a906-f91e1432adbf\") " pod="openshift-debug-4sl8v/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl-debug" Jan 05 10:02:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 10:02:50.628373 1977 operation_generator.go:730] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcwnr\" (UniqueName: \"kubernetes.io/projected/281d0fef-ebf1-4d7c-a906-f91e1432adbf-kube-api-access-gcwnr\") pod \"pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl-debug\" (UID: \"281d0fef-ebf1-4d7c-a906-f91e1432adbf\") " pod="openshift-debug-4sl8v/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl-debug" Jan 05 10:02:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 10:02:50.666057 1977 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-debug-4sl8v/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl-debug" Jan 05 10:02:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 10:02:50.666537867Z" level=info msg="Running pod sandbox: openshift-debug-4sl8v/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl-debug/POD" id=0d73d7d8-7db9-4cce-810e-94107a03c047 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 10:02:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 10:02:50.666595515Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 10:02:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 10:02:50.680589716Z" level=warning msg="Skipping invalid sysctl specified by config {net.ipv4.ping_group_range 0 2147483647}: \"net.ipv4.ping_group_range\" not allowed with host net enabled" id=0d73d7d8-7db9-4cce-810e-94107a03c047 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 10:02:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: W0105 10:02:50.686608 1977 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod281d0fef_ebf1_4d7c_a906_f91e1432adbf.slice/crio-580d06b1ae2293d1bcfb3d177e8d2383ed0b9e8eb15b5da68f93d70ede36aa7f.scope WatchSource:0}: Error finding container 580d06b1ae2293d1bcfb3d177e8d2383ed0b9e8eb15b5da68f93d70ede36aa7f: Status 404 returned error can't find the container with id 580d06b1ae2293d1bcfb3d177e8d2383ed0b9e8eb15b5da68f93d70ede36aa7f Jan 05 10:02:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 10:02:50.700685809Z" level=info msg="Ran pod sandbox 580d06b1ae2293d1bcfb3d177e8d2383ed0b9e8eb15b5da68f93d70ede36aa7f with infra container: openshift-debug-4sl8v/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl-debug/POD" id=0d73d7d8-7db9-4cce-810e-94107a03c047 name=/runtime.v1.RuntimeService/RunPodSandbox Jan 05 10:02:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 10:02:50.701959426Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b1ede5afff16c2d6d9cbfafc77bb0aea9f76fc79727dc365be5d8ee908158da0" id=b434f331-47b9-4f40-a41a-cb1026054836 name=/runtime.v1.ImageService/ImageStatus Jan 05 10:02:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 10:02:50.702146730Z" level=info msg="Image registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b1ede5afff16c2d6d9cbfafc77bb0aea9f76fc79727dc365be5d8ee908158da0 not found" id=b434f331-47b9-4f40-a41a-cb1026054836 name=/runtime.v1.ImageService/ImageStatus Jan 05 10:02:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 10:02:50.704438 1977 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 05 10:02:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 10:02:50.705082996Z" level=info msg="Pulling image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b1ede5afff16c2d6d9cbfafc77bb0aea9f76fc79727dc365be5d8ee908158da0" id=4ea5ff28-904c-4337-b5a4-50cee4fad74e name=/runtime.v1.ImageService/PullImage Jan 05 10:02:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 10:02:50.727254122Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b1ede5afff16c2d6d9cbfafc77bb0aea9f76fc79727dc365be5d8ee908158da0\"" Jan 05 10:02:50 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 10:02:50.936049248Z" level=info msg="Trying to access \"registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b1ede5afff16c2d6d9cbfafc77bb0aea9f76fc79727dc365be5d8ee908158da0\"" Jan 05 10:02:51 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 10:02:51.080644 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-debug-4sl8v/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl-debug" event=&{ID:281d0fef-ebf1-4d7c-a906-f91e1432adbf Type:ContainerStarted Data:580d06b1ae2293d1bcfb3d177e8d2383ed0b9e8eb15b5da68f93d70ede36aa7f} Jan 05 10:03:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 10:03:02.584229713Z" level=info msg="Pulled image: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b1ede5afff16c2d6d9cbfafc77bb0aea9f76fc79727dc365be5d8ee908158da0" id=4ea5ff28-904c-4337-b5a4-50cee4fad74e name=/runtime.v1.ImageService/PullImage Jan 05 10:03:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 10:03:02.585609358Z" level=info msg="Checking image status: registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b1ede5afff16c2d6d9cbfafc77bb0aea9f76fc79727dc365be5d8ee908158da0" id=f9765604-b2f3-4b9d-a832-c936baaba655 name=/runtime.v1.ImageService/ImageStatus Jan 05 10:03:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 10:03:02.587286168Z" level=info msg="Image status: &ImageStatusResponse{Image:&Image{Id:372d2208073802051a8f01c968360ff91a83a269588e195e94db9b4029966775,RepoTags:[],RepoDigests:[registry.ci.openshift.org/ocp/4.12-2023-01-05-021459@sha256:b1ede5afff16c2d6d9cbfafc77bb0aea9f76fc79727dc365be5d8ee908158da0],Size_:780064028,Uid:&Int64Value{Value:0,},Username:,Spec:nil,Pinned:false,},Info:map[string]string{},}" id=f9765604-b2f3-4b9d-a832-c936baaba655 name=/runtime.v1.ImageService/ImageStatus Jan 05 10:03:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 10:03:02.587968118Z" level=info msg="Creating container: openshift-debug-4sl8v/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl-debug/container-00" id=086db084-9602-4ec0-8721-3a8cb4b4ab40 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 10:03:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 10:03:02.588074665Z" level=warning msg="Allowed annotations are specified for workload [io.containers.trace-syscall]" Jan 05 10:03:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 10:03:02.687477705Z" level=info msg="Created container 7b6b49a262d6b42611e84a284632662a22f5c87f98083a6aec12b82efd3f20d9: openshift-debug-4sl8v/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl-debug/container-00" id=086db084-9602-4ec0-8721-3a8cb4b4ab40 name=/runtime.v1.RuntimeService/CreateContainer Jan 05 10:03:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 10:03:02.688140421Z" level=info msg="Starting container: 7b6b49a262d6b42611e84a284632662a22f5c87f98083a6aec12b82efd3f20d9" id=e99257cf-5dde-46f5-84ec-292babd3181b name=/runtime.v1.RuntimeService/StartContainer Jan 05 10:03:02 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl crio[1941]: time="2023-01-05 10:03:02.697186336Z" level=info msg="Started container" PID=23154 containerID=7b6b49a262d6b42611e84a284632662a22f5c87f98083a6aec12b82efd3f20d9 description=openshift-debug-4sl8v/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl-debug/container-00 id=e99257cf-5dde-46f5-84ec-292babd3181b name=/runtime.v1.RuntimeService/StartContainer sandboxID=580d06b1ae2293d1bcfb3d177e8d2383ed0b9e8eb15b5da68f93d70ede36aa7f Jan 05 10:03:03 pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl kubenswrapper[1977]: I0105 10:03:03.139254 1977 kubelet.go:2157] "SyncLoop (PLEG): event for pod" pod="openshift-debug-4sl8v/pliu-dev-alicloud-fp8fj-worker-us-east-1a-nm7pl-debug" event=&{ID:281d0fef-ebf1-4d7c-a906-f91e1432adbf Type:ContainerStarted Data:7b6b49a262d6b42611e84a284632662a22f5c87f98083a6aec12b82efd3f20d9} sh-4.4#