$ hack/local-e2e-test.sh 2026-02-02T08:37:35,921429961+00:00 Running e2e with HIVE_IMAGE quay.io/redhat-user-workloads/crt-redhat-acm-tenant/hive-operator/hive:hive-on-push-9v9lm-build-image-index 2026-02-02T08:37:35,924334578+00:00 Running e2e with RELEASE_IMAGE quay.io/openshift-release-dev/ocp-release:4.20.13-x86_64 Error from server (NotFound): namespaces "hive-e2e1" not found 2026-02-02T08:37:36,062716298+00:00 Creating namespace hive-e2e1. Try #1/120... namespace/hive-e2e1 created 2026-02-02T08:37:36,190597064+00:00 Success 2026-02-02T08:37:36,193543269+00:00 Setting default namespace to hive-e2e1 Context "admin" modified. Defaulting to user installation because normal site-packages is not writeable Looking in links: /tmp/tmp_28th1rd Requirement already satisfied: setuptools in /usr/lib/python3.9/site-packages (53.0.0) Requirement already satisfied: pip in /home/cloud-user/.local/lib/python3.9/site-packages (25.3) Requirement already satisfied: gitpython in /home/cloud-user/.local/lib/python3.9/site-packages (3.1.45) Requirement already satisfied: pyyaml in /usr/lib64/python3.9/site-packages (5.4.1) Requirement already satisfied: gitdb<5,>=4.0.1 in /home/cloud-user/.local/lib/python3.9/site-packages (from gitpython) (4.0.12) Requirement already satisfied: typing-extensions>=3.10.0.2 in /home/cloud-user/.local/lib/python3.9/site-packages (from gitpython) (4.14.1) Requirement already satisfied: smmap<6,>=3.0.1 in /home/cloud-user/.local/lib/python3.9/site-packages (from gitdb<5,>=4.0.1->gitpython) (5.0.2) [notice] A new release of pip is available: 25.3 -> 26.0 [notice] To update, run: python3 -m pip install --upgrade pip Using active branch HIVE-3067 since it corresponds to commit b8f7293 Using existing kustomize from "_output/tools/bin/kustomize-4.1.3" Using existing yq from "_output/tools/bin/yq-v4.47.1" rm -rf ./config/crds (cd apis; /home/cloud-user/go/bin/controller-gen crd:crdVersions=v1 paths=./hive/v1 paths=./hiveinternal/v1alpha1 output:dir=../config/crds) Stripping yaml breaks from CRD files _output/tools/bin/yq-v4.47.1 --no-doc -i ./config/crds/hive.openshift.io_checkpoints.yaml; _output/tools/bin/yq-v4.47.1 --no-doc -i ./config/crds/hive.openshift.io_clusterclaims.yaml; _output/tools/bin/yq-v4.47.1 --no-doc -i ./config/crds/hive.openshift.io_clusterdeploymentcustomizations.yaml; _output/tools/bin/yq-v4.47.1 --no-doc -i ./config/crds/hive.openshift.io_clusterdeployments.yaml; _output/tools/bin/yq-v4.47.1 --no-doc -i ./config/crds/hive.openshift.io_clusterdeprovisions.yaml; _output/tools/bin/yq-v4.47.1 --no-doc -i ./config/crds/hive.openshift.io_clusterimagesets.yaml; _output/tools/bin/yq-v4.47.1 --no-doc -i ./config/crds/hive.openshift.io_clusterpools.yaml; _output/tools/bin/yq-v4.47.1 --no-doc -i ./config/crds/hive.openshift.io_clusterprovisions.yaml; _output/tools/bin/yq-v4.47.1 --no-doc -i ./config/crds/hive.openshift.io_clusterrelocates.yaml; _output/tools/bin/yq-v4.47.1 --no-doc -i ./config/crds/hive.openshift.io_clusterstates.yaml; _output/tools/bin/yq-v4.47.1 --no-doc -i ./config/crds/hive.openshift.io_dnszones.yaml; _output/tools/bin/yq-v4.47.1 --no-doc -i ./config/crds/hive.openshift.io_hiveconfigs.yaml; _output/tools/bin/yq-v4.47.1 --no-doc -i ./config/crds/hive.openshift.io_machinepoolnameleases.yaml; _output/tools/bin/yq-v4.47.1 --no-doc -i ./config/crds/hive.openshift.io_machinepools.yaml; _output/tools/bin/yq-v4.47.1 --no-doc -i ./config/crds/hive.openshift.io_selectorsyncidentityproviders.yaml; _output/tools/bin/yq-v4.47.1 --no-doc -i ./config/crds/hive.openshift.io_selectorsyncsets.yaml; _output/tools/bin/yq-v4.47.1 --no-doc -i ./config/crds/hive.openshift.io_syncidentityproviders.yaml; _output/tools/bin/yq-v4.47.1 --no-doc -i ./config/crds/hive.openshift.io_syncsets.yaml; _output/tools/bin/yq-v4.47.1 --no-doc -i ./config/crds/hiveinternal.openshift.io_clustersyncleases.yaml; _output/tools/bin/yq-v4.47.1 --no-doc -i ./config/crds/hiveinternal.openshift.io_clustersyncs.yaml; _output/tools/bin/yq-v4.47.1 --no-doc -i ./config/crds/hiveinternal.openshift.io_fakeclusterinstalls.yaml; Patching CRD files for additional static information _output/tools/bin/yq-v4.47.1 '. * load("./config/crdspatch/hive.openshift.io_clusterprovisions.yaml")' -i './config/crds/hive.openshift.io_clusterprovisions.yaml'; _output/tools/bin/yq-v4.47.1 '. * load("./config/crdspatch/hiveinternal.openshift.io_fakeclusterinstalls.yaml")' -i './config/crds/hiveinternal.openshift.io_fakeclusterinstalls.yaml'; # Patch ClusterProvision CRD to remove the massive PodSpec def we consider an internal implementation detail: Patching ClusterProvision CRD yaml to remove overly verbose PodSpec details: _output/tools/bin/yq-v4.47.1 eval 'del(.spec.versions[0].schema.openAPIV3Schema.properties.spec.properties.podSpec)' -i config/crds/hive.openshift.io_clusterprovisions.yaml # This does not appear possible with controller-runtime flags when dealing with an array, # kubebuilder:validation:EmbeddedResource adds the x-kubernetes-embedded-resource to the array, # not the elements within it. Patching SyncSet CRDs to flag resource RawExtensions as embedded resources: _output/tools/bin/yq-v4.47.1 eval '.spec.versions[0].schema.openAPIV3Schema.properties.spec.properties.resources.items.x-kubernetes-embedded-resource = true' -i config/crds/hive.openshift.io_syncsets.yaml _output/tools/bin/yq-v4.47.1 eval '.spec.versions[0].schema.openAPIV3Schema.properties.spec.properties.resources.items.x-kubernetes-preserve-unknown-fields = true' -i config/crds/hive.openshift.io_syncsets.yaml _output/tools/bin/yq-v4.47.1 eval '.spec.versions[0].schema.openAPIV3Schema.properties.spec.properties.resources.items.x-kubernetes-embedded-resource = true' -i config/crds/hive.openshift.io_selectorsyncsets.yaml _output/tools/bin/yq-v4.47.1 eval '.spec.versions[0].schema.openAPIV3Schema.properties.spec.properties.resources.items.x-kubernetes-preserve-unknown-fields = true' -i config/crds/hive.openshift.io_selectorsyncsets.yaml oc apply -f config/crds customresourcedefinition.apiextensions.k8s.io/checkpoints.hive.openshift.io unchanged customresourcedefinition.apiextensions.k8s.io/clusterclaims.hive.openshift.io unchanged customresourcedefinition.apiextensions.k8s.io/clusterdeploymentcustomizations.hive.openshift.io unchanged customresourcedefinition.apiextensions.k8s.io/clusterdeployments.hive.openshift.io unchanged customresourcedefinition.apiextensions.k8s.io/clusterdeprovisions.hive.openshift.io unchanged customresourcedefinition.apiextensions.k8s.io/clusterimagesets.hive.openshift.io unchanged customresourcedefinition.apiextensions.k8s.io/clusterpools.hive.openshift.io unchanged customresourcedefinition.apiextensions.k8s.io/clusterprovisions.hive.openshift.io unchanged customresourcedefinition.apiextensions.k8s.io/clusterrelocates.hive.openshift.io unchanged customresourcedefinition.apiextensions.k8s.io/clusterstates.hive.openshift.io unchanged customresourcedefinition.apiextensions.k8s.io/dnszones.hive.openshift.io unchanged customresourcedefinition.apiextensions.k8s.io/hiveconfigs.hive.openshift.io unchanged customresourcedefinition.apiextensions.k8s.io/machinepoolnameleases.hive.openshift.io unchanged customresourcedefinition.apiextensions.k8s.io/machinepools.hive.openshift.io unchanged customresourcedefinition.apiextensions.k8s.io/selectorsyncidentityproviders.hive.openshift.io unchanged customresourcedefinition.apiextensions.k8s.io/selectorsyncsets.hive.openshift.io unchanged customresourcedefinition.apiextensions.k8s.io/syncidentityproviders.hive.openshift.io unchanged customresourcedefinition.apiextensions.k8s.io/syncsets.hive.openshift.io unchanged customresourcedefinition.apiextensions.k8s.io/clustersyncleases.hiveinternal.openshift.io unchanged customresourcedefinition.apiextensions.k8s.io/clustersyncs.hiveinternal.openshift.io unchanged customresourcedefinition.apiextensions.k8s.io/fakeclusterinstalls.hiveinternal.openshift.io unchanged # Deploy the operator manifests: oc create namespace hive-operator || true namespace/hive-operator created mkdir -p overlays/deploy cp overlays/template/kustomization.yaml overlays/deploy cd overlays/deploy && ../../_output/tools/bin/kustomize-4.1.3 edit set image registry.ci.openshift.org/openshift/hive-v4.0:hive=quay.io/redhat-user-workloads/crt-redhat-acm-tenant/hive-operator/hive:hive-on-push-9v9lm-build-image-index && ../../_output/tools/bin/kustomize-4.1.3 edit set namespace hive-operator _output/tools/bin/kustomize-4.1.3 build overlays/deploy | sed 's/ - info/ - debug/' | oc apply -f - serviceaccount/hive-operator created clusterrole.rbac.authorization.k8s.io/hive-operator-role created clusterrolebinding.rbac.authorization.k8s.io/hive-operator-rolebinding created deployment.apps/hive-operator created rm -rf overlays/deploy # Create a default basic HiveConfig so the operator will deploy Hive oc process --local=true -p HIVE_NS=hive-e2e -p LOG_LEVEL=debug -f config/templates/hiveconfig.yaml | oc apply -f - hiveconfig.hive.openshift.io/hive created 2026-02-02T08:37:42,281741000+00:00 Waiting 180 seconds for namespace hive-e2e to appear2026-02-02T08:37:42,420802086+00:00 .2026-02-02T08:37:43,582013836+00:00 .2026-02-02T08:37:44,722277536+00:00 .2026-02-02T08:37:45,875791077+00:00 2026-02-02T08:37:45,882111373+00:00 Using cluster base domain: qe.devcluster.openshift.com No resources found in hive-e2e1 namespace. 2026-02-02T08:37:46,016126155+00:00 Running post-deploy tests in original namespace hive-e2e Defaulting to user installation because normal site-packages is not writeable Looking in links: /tmp/tmp0smdef38 Requirement already satisfied: setuptools in /usr/lib/python3.9/site-packages (53.0.0) Requirement already satisfied: pip in /home/cloud-user/.local/lib/python3.9/site-packages (25.3) Requirement already satisfied: gitpython in /home/cloud-user/.local/lib/python3.9/site-packages (3.1.45) Requirement already satisfied: pyyaml in /usr/lib64/python3.9/site-packages (5.4.1) Requirement already satisfied: gitdb<5,>=4.0.1 in /home/cloud-user/.local/lib/python3.9/site-packages (from gitpython) (4.0.12) Requirement already satisfied: typing-extensions>=3.10.0.2 in /home/cloud-user/.local/lib/python3.9/site-packages (from gitpython) (4.14.1) Requirement already satisfied: smmap<6,>=3.0.1 in /home/cloud-user/.local/lib/python3.9/site-packages (from gitdb<5,>=4.0.1->gitpython) (5.0.2) [notice] A new release of pip is available: 25.3 -> 26.0 [notice] To update, run: python3 -m pip install --upgrade pip Using active branch HIVE-3067 since it corresponds to commit b8f7293 go test -mod=vendor -v -timeout 0 -count=1 ./test/e2e/postdeploy/... === RUN TestAdmission time="2026-02-02T03:37:51-05:00" level=info msg="Registering validation REST resource" group=admission.hive.openshift.io resource=dnszonevalidator version=v1 time="2026-02-02T03:37:51-05:00" level=info msg="Waiting for deployment" deployment=hive-e2e/hiveadmission time="2026-02-02T03:37:51-05:00" level=info msg="Waiting for APIService" APIService=v1.admission.hive.openshift.io === RUN TestAdmission/review_failure I0202 03:37:51.806568 3063020 warnings.go:110] "Warning: admission.k8s.io/v1beta1 AdmissionReview is deprecated in v1.19+, unavailable in v1.22+; use admission.k8s.io/v1 AdmissionReview" === RUN TestAdmission/review_success I0202 03:37:51.818881 3063020 warnings.go:110] "Warning: admission.k8s.io/v1beta1 AdmissionReview is deprecated in v1.19+, unavailable in v1.22+; use admission.k8s.io/v1 AdmissionReview" === RUN TestAdmission/review_update_failure I0202 03:37:51.829703 3063020 warnings.go:110] "Warning: admission.k8s.io/v1beta1 AdmissionReview is deprecated in v1.19+, unavailable in v1.22+; use admission.k8s.io/v1 AdmissionReview" --- PASS: TestAdmission (0.72s) --- PASS: TestAdmission/review_failure (0.02s) --- PASS: TestAdmission/review_success (0.01s) --- PASS: TestAdmission/review_update_failure (0.01s) PASS ok github.com/openshift/hive/test/e2e/postdeploy/admission 0.764s === RUN TestHiveControllersDeployment time="2026-02-02T03:37:51-05:00" level=info msg="Waiting for deployment" deployment=hive-e2e/hive-controllers --- PASS: TestHiveControllersDeployment (0.07s) === RUN TestHiveControllersMetrics time="2026-02-02T03:37:51-05:00" level=info msg="Waiting for deployment" deployment=hive-e2e/hive-controllers --- PASS: TestHiveControllersMetrics (0.04s) PASS ok github.com/openshift/hive/test/e2e/postdeploy/hivecontroller 0.161s === RUN TestOperatorDeployment time="2026-02-02T03:37:51-05:00" level=info msg="Waiting for deployment" deployment=hive-operator/hive-operator --- PASS: TestOperatorDeployment (0.07s) === RUN TestHiveCRDs --- PASS: TestHiveCRDs (0.39s) === RUN TestHiveConfig --- PASS: TestHiveConfig (0.02s) PASS ok github.com/openshift/hive/test/e2e/postdeploy/operator 0.557s hiveconfig.hive.openshift.io/hive patched 2026-02-02T08:37:52,013793328+00:00 Waiting 60 seconds for namespace hive-e2e-two to appear2026-02-02T08:37:52,151485399+00:00 .2026-02-02T08:37:53,300699266+00:00 Running post-deploy tests in new namespace hive-e2e-two Defaulting to user installation because normal site-packages is not writeable Looking in links: /tmp/tmpevfbzdmg Requirement already satisfied: setuptools in /usr/lib/python3.9/site-packages (53.0.0) Requirement already satisfied: pip in /home/cloud-user/.local/lib/python3.9/site-packages (25.3) Requirement already satisfied: gitpython in /home/cloud-user/.local/lib/python3.9/site-packages (3.1.45) Requirement already satisfied: pyyaml in /usr/lib64/python3.9/site-packages (5.4.1) Requirement already satisfied: gitdb<5,>=4.0.1 in /home/cloud-user/.local/lib/python3.9/site-packages (from gitpython) (4.0.12) Requirement already satisfied: typing-extensions>=3.10.0.2 in /home/cloud-user/.local/lib/python3.9/site-packages (from gitpython) (4.14.1) Requirement already satisfied: smmap<6,>=3.0.1 in /home/cloud-user/.local/lib/python3.9/site-packages (from gitdb<5,>=4.0.1->gitpython) (5.0.2) [notice] A new release of pip is available: 25.3 -> 26.0 [notice] To update, run: python3 -m pip install --upgrade pip Using active branch HIVE-3067 since it corresponds to commit b8f7293 go test -mod=vendor -v -timeout 0 -count=1 ./test/e2e/postdeploy/... === RUN TestAdmission time="2026-02-02T03:37:58-05:00" level=info msg="Registering validation REST resource" group=admission.hive.openshift.io resource=dnszonevalidator version=v1 time="2026-02-02T03:37:58-05:00" level=info msg="Waiting for deployment" deployment=hive-e2e-two/hiveadmission time="2026-02-02T03:37:59-05:00" level=info msg="Waiting for APIService" APIService=v1.admission.hive.openshift.io === RUN TestAdmission/review_failure I0202 03:37:59.874616 3063541 warnings.go:110] "Warning: admission.k8s.io/v1beta1 AdmissionReview is deprecated in v1.19+, unavailable in v1.22+; use admission.k8s.io/v1 AdmissionReview" === RUN TestAdmission/review_success I0202 03:37:59.890462 3063541 warnings.go:110] "Warning: admission.k8s.io/v1beta1 AdmissionReview is deprecated in v1.19+, unavailable in v1.22+; use admission.k8s.io/v1 AdmissionReview" === RUN TestAdmission/review_update_failure I0202 03:37:59.901852 3063541 warnings.go:110] "Warning: admission.k8s.io/v1beta1 AdmissionReview is deprecated in v1.19+, unavailable in v1.22+; use admission.k8s.io/v1 AdmissionReview" --- PASS: TestAdmission (1.58s) --- PASS: TestAdmission/review_failure (0.01s) --- PASS: TestAdmission/review_success (0.02s) --- PASS: TestAdmission/review_update_failure (0.01s) PASS ok github.com/openshift/hive/test/e2e/postdeploy/admission 1.639s === RUN TestHiveControllersDeployment time="2026-02-02T03:37:58-05:00" level=info msg="Waiting for deployment" deployment=hive-e2e-two/hive-controllers --- PASS: TestHiveControllersDeployment (0.07s) === RUN TestHiveControllersMetrics time="2026-02-02T03:37:58-05:00" level=info msg="Waiting for deployment" deployment=hive-e2e-two/hive-controllers --- PASS: TestHiveControllersMetrics (0.03s) PASS ok github.com/openshift/hive/test/e2e/postdeploy/hivecontroller 0.156s === RUN TestOperatorDeployment time="2026-02-02T03:37:58-05:00" level=info msg="Waiting for deployment" deployment=hive-operator/hive-operator --- PASS: TestOperatorDeployment (0.07s) === RUN TestHiveCRDs --- PASS: TestHiveCRDs (0.39s) === RUN TestHiveConfig --- PASS: TestHiveConfig (0.02s) PASS ok github.com/openshift/hive/test/e2e/postdeploy/operator 0.537s 2026-02-02T08:37:59,940875730+00:00 Checking for stale secret resources in original namespace hive-e2e 2026-02-02T08:38:00,071154219+00:00 Checking for stale configmap resources in original namespace hive-e2e 2026-02-02T08:38:00,208515512+00:00 Checking for stale role resources in original namespace hive-e2e No resources found in hive-e2e namespace. 2026-02-02T08:38:00,344896220+00:00 Checking for stale rolebinding resources in original namespace hive-e2e 2026-02-02T08:38:00,477639116+00:00 Checking for stale serviceaccount resources in original namespace hive-e2e 2026-02-02T08:38:00,622006546+00:00 Checking for stale deployment resources in original namespace hive-e2e No resources found in hive-e2e namespace. 2026-02-02T08:38:00,753018659+00:00 Checking for stale replicaset resources in original namespace hive-e2e No resources found in hive-e2e namespace. 2026-02-02T08:38:00,887703101+00:00 Checking for stale statefulset resources in original namespace hive-e2e No resources found in hive-e2e namespace. 2026-02-02T08:38:01,018374840+00:00 Checking for stale pod resources in original namespace hive-e2e No resources found in hive-e2e namespace. 2026-02-02T08:38:01,155866459+00:00 Creating cluster deployment DEBU[0000] Couldn't find install logs provider environment variable. Skipping. clusterdeployment.hive.openshift.io/hive-7dda6e0a-a27e-4845-9609-3814aaddac46 created machinepool.hive.openshift.io/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker created secret/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-install-config created secret/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-pull-secret created secret/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-aws-creds created clusterimageset.hive.openshift.io/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-imageset created 2026-02-02T08:38:02,864058502+00:00 Creating fake clusterinstall artifacts clusterimageset.hive.openshift.io/hive-not-a-release-image unchanged clusterdeployment.hive.openshift.io/fci created fakeclusterinstall.hiveinternal.openshift.io/fci created 2026-02-02T08:38:03,475016633+00:00 Getting ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46. Try #1/120: 2026-02-02T08:38:03,621398332+00:00 NAME INFRAID PLATFORM REGION VERSION CLUSTERTYPE PROVISIONSTATUS POWERSTATE AGE 2026-02-02T08:38:03,624724078+00:00 Success 2026-02-02T08:40:03,628976809+00:00 Deployments in hive namespace NAME READY UP-TO-DATE AVAILABLE AGE hive-controllers 1/1 1 1 2m10s hiveadmission 2/2 2 2 2m6s 2026-02-02T08:40:03,791155352+00:00 2026-02-02T08:40:03,793867516+00:00 Pods in hive namespace NAME READY STATUS RESTARTS AGE hive-clustersync-0 1/1 Running 0 2m10s hive-controllers-7d4b854fb9-zmwb8 1/1 Running 0 2m10s hive-machinepool-0 1/1 Running 0 2m10s hiveadmission-78b7cd97bb-4kffx 1/1 Running 0 2m6s hiveadmission-78b7cd97bb-zknwx 1/1 Running 0 2m6s 2026-02-02T08:40:03,935442818+00:00 2026-02-02T08:40:03,938469389+00:00 Pods in cluster namespace NAME READY STATUS RESTARTS AGE hive-7dda6e0a-a27e-4845-9609-3814aaddac46-0-hcbfb-provisioctt22 1/1 Running 0 111s 2026-02-02T08:40:04,079984018+00:00 2026-02-02T08:40:04,083035280+00:00 Events in hive namespace LAST SEEN TYPE REASON OBJECT MESSAGE 2m9s Normal Scheduled pod/hive-clustersync-0 Successfully assigned hive-e2e-two/hive-clustersync-0 to ip-10-0-62-221.ec2.internal 2m9s Normal AddedInterface pod/hive-clustersync-0 Add eth0 [10.128.2.41/23] from ovn-kubernetes 2m9s Normal Pulling pod/hive-clustersync-0 Pulling image "quay.io/redhat-user-workloads/crt-redhat-acm-tenant/hive-operator/hive:hive-on-push-9v9lm-build-image-index" 2m9s Normal Pulled pod/hive-clustersync-0 Successfully pulled image "quay.io/redhat-user-workloads/crt-redhat-acm-tenant/hive-operator/hive:hive-on-push-9v9lm-build-image-index" in 309ms (309ms including waiting). Image size: 1938808453 bytes. 2m9s Normal Created pod/hive-clustersync-0 Created container: clustersync 2m9s Normal Started pod/hive-clustersync-0 Started container clustersync 2m10s Normal SuccessfulCreate statefulset/hive-clustersync create Pod hive-clustersync-0 in StatefulSet hive-clustersync successful 2m10s Normal Scheduled pod/hive-controllers-7d4b854fb9-zmwb8 Successfully assigned hive-e2e-two/hive-controllers-7d4b854fb9-zmwb8 to ip-10-0-38-164.ec2.internal 2m9s Normal AddedInterface pod/hive-controllers-7d4b854fb9-zmwb8 Add eth0 [10.129.2.101/23] from ovn-kubernetes 2m9s Normal Pulling pod/hive-controllers-7d4b854fb9-zmwb8 Pulling image "quay.io/redhat-user-workloads/crt-redhat-acm-tenant/hive-operator/hive:hive-on-push-9v9lm-build-image-index" 2m9s Normal Pulled pod/hive-controllers-7d4b854fb9-zmwb8 Successfully pulled image "quay.io/redhat-user-workloads/crt-redhat-acm-tenant/hive-operator/hive:hive-on-push-9v9lm-build-image-index" in 397ms (397ms including waiting). Image size: 1938808453 bytes. 2m9s Normal Created pod/hive-controllers-7d4b854fb9-zmwb8 Created container: manager 2m9s Normal Started pod/hive-controllers-7d4b854fb9-zmwb8 Started container manager 2m10s Normal SuccessfulCreate replicaset/hive-controllers-7d4b854fb9 Created pod: hive-controllers-7d4b854fb9-zmwb8 2m10s Normal ScalingReplicaSet deployment/hive-controllers Scaled up replica set hive-controllers-7d4b854fb9 from 0 to 1 2m9s Normal Scheduled pod/hive-machinepool-0 Successfully assigned hive-e2e-two/hive-machinepool-0 to ip-10-0-38-164.ec2.internal 2m9s Normal AddedInterface pod/hive-machinepool-0 Add eth0 [10.129.2.102/23] from ovn-kubernetes 2m9s Normal Pulling pod/hive-machinepool-0 Pulling image "quay.io/redhat-user-workloads/crt-redhat-acm-tenant/hive-operator/hive:hive-on-push-9v9lm-build-image-index" 2m9s Normal Pulled pod/hive-machinepool-0 Successfully pulled image "quay.io/redhat-user-workloads/crt-redhat-acm-tenant/hive-operator/hive:hive-on-push-9v9lm-build-image-index" in 396ms (396ms including waiting). Image size: 1938808453 bytes. 2m9s Normal Created pod/hive-machinepool-0 Created container: machinepool 2m9s Normal Started pod/hive-machinepool-0 Started container machinepool 2m10s Normal SuccessfulCreate statefulset/hive-machinepool create Pod hive-machinepool-0 in StatefulSet hive-machinepool successful 2m6s Normal Scheduled pod/hiveadmission-78b7cd97bb-4kffx Successfully assigned hive-e2e-two/hiveadmission-78b7cd97bb-4kffx to ip-10-0-38-164.ec2.internal 2m6s Normal AddedInterface pod/hiveadmission-78b7cd97bb-4kffx Add eth0 [10.129.2.103/23] from ovn-kubernetes 2m6s Normal Pulling pod/hiveadmission-78b7cd97bb-4kffx Pulling image "quay.io/redhat-user-workloads/crt-redhat-acm-tenant/hive-operator/hive:hive-on-push-9v9lm-build-image-index" 2m5s Normal Pulled pod/hiveadmission-78b7cd97bb-4kffx Successfully pulled image "quay.io/redhat-user-workloads/crt-redhat-acm-tenant/hive-operator/hive:hive-on-push-9v9lm-build-image-index" in 344ms (344ms including waiting). Image size: 1938808453 bytes. 2m5s Normal Created pod/hiveadmission-78b7cd97bb-4kffx Created container: hiveadmission 2m5s Normal Started pod/hiveadmission-78b7cd97bb-4kffx Started container hiveadmission 2m6s Normal Scheduled pod/hiveadmission-78b7cd97bb-zknwx Successfully assigned hive-e2e-two/hiveadmission-78b7cd97bb-zknwx to ip-10-0-74-176.ec2.internal 2m6s Normal AddedInterface pod/hiveadmission-78b7cd97bb-zknwx Add eth0 [10.131.0.89/23] from ovn-kubernetes 2m6s Normal Pulling pod/hiveadmission-78b7cd97bb-zknwx Pulling image "quay.io/redhat-user-workloads/crt-redhat-acm-tenant/hive-operator/hive:hive-on-push-9v9lm-build-image-index" 2m5s Normal Pulled pod/hiveadmission-78b7cd97bb-zknwx Successfully pulled image "quay.io/redhat-user-workloads/crt-redhat-acm-tenant/hive-operator/hive:hive-on-push-9v9lm-build-image-index" in 316ms (316ms including waiting). Image size: 1938808453 bytes. 2m5s Normal Created pod/hiveadmission-78b7cd97bb-zknwx Created container: hiveadmission 2m5s Normal Started pod/hiveadmission-78b7cd97bb-zknwx Started container hiveadmission 2m6s Normal SuccessfulCreate replicaset/hiveadmission-78b7cd97bb Created pod: hiveadmission-78b7cd97bb-4kffx 2m6s Normal SuccessfulCreate replicaset/hiveadmission-78b7cd97bb Created pod: hiveadmission-78b7cd97bb-zknwx 2m6s Normal ScalingReplicaSet deployment/hiveadmission Scaled up replica set hiveadmission-78b7cd97bb from 0 to 2 2026-02-02T08:40:04,232360887+00:00 2026-02-02T08:40:04,235682078+00:00 Events in cluster namespace LAST SEEN TYPE REASON OBJECT MESSAGE 111s Normal Scheduled pod/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-0-hcbfb-provisioctt22 Successfully assigned hive-e2e1/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-0-hcbfb-provisioctt22 to ip-10-0-38-164.ec2.internal 110s Normal AddedInterface pod/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-0-hcbfb-provisioctt22 Add eth0 [10.129.2.105/23] from ovn-kubernetes 110s Normal Pulling pod/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-0-hcbfb-provisioctt22 Pulling image "quay.io/redhat-user-workloads/crt-redhat-acm-tenant/hive-operator/hive:hive-on-push-9v9lm-build-image-index" 110s Normal Pulled pod/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-0-hcbfb-provisioctt22 Successfully pulled image "quay.io/redhat-user-workloads/crt-redhat-acm-tenant/hive-operator/hive:hive-on-push-9v9lm-build-image-index" in 304ms (304ms including waiting). Image size: 1938808453 bytes. 110s Normal Created pod/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-0-hcbfb-provisioctt22 Created container: hive 110s Normal Started pod/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-0-hcbfb-provisioctt22 Started container hive 109s Normal Pulling pod/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-0-hcbfb-provisioctt22 Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:80819f88795a16b1df3cea66b7fe57d68b504b08fcbaaf245716f3caed588e82" 107s Normal Pulled pod/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-0-hcbfb-provisioctt22 Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:80819f88795a16b1df3cea66b7fe57d68b504b08fcbaaf245716f3caed588e82" in 2.102s (2.102s including waiting). Image size: 590116485 bytes. 107s Normal Created pod/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-0-hcbfb-provisioctt22 Created container: cli 107s Normal Started pod/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-0-hcbfb-provisioctt22 Started container cli 106s Normal Pulling pod/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-0-hcbfb-provisioctt22 Pulling image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:209f7fa1c2d11b79137aeeb1124d6bba014b76163d04ef8b46fd3662cd85720d" 89s Normal Pulled pod/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-0-hcbfb-provisioctt22 Successfully pulled image "quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:209f7fa1c2d11b79137aeeb1124d6bba014b76163d04ef8b46fd3662cd85720d" in 17.906s (17.906s including waiting). Image size: 1591835361 bytes. 88s Normal Created pod/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-0-hcbfb-provisioctt22 Created container: installer 88s Normal Started pod/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-0-hcbfb-provisioctt22 Started container installer 111s Normal SuccessfulCreate job/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-0-hcbfb-provision Created pod: hive-7dda6e0a-a27e-4845-9609-3814aaddac46-0-hcbfb-provisioctt22 119s Normal Scheduled pod/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-imageset-8tc8g Successfully assigned hive-e2e1/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-imageset-8tc8g to ip-10-0-38-164.ec2.internal 119s Normal AddedInterface pod/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-imageset-8tc8g Add eth0 [10.129.2.104/23] from ovn-kubernetes 119s Normal Pulling pod/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-imageset-8tc8g Pulling image "quay.io/openshift-release-dev/ocp-release:4.20.13-x86_64" 115s Normal Pulled pod/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-imageset-8tc8g Successfully pulled image "quay.io/openshift-release-dev/ocp-release:4.20.13-x86_64" in 3.61s (3.61s including waiting). Image size: 503650873 bytes. 115s Normal Created pod/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-imageset-8tc8g Created container: release 115s Normal Started pod/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-imageset-8tc8g Started container release 114s Normal Pulling pod/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-imageset-8tc8g Pulling image "quay.io/redhat-user-workloads/crt-redhat-acm-tenant/hive-operator/hive:hive-on-push-9v9lm-build-image-index" 114s Normal Pulled pod/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-imageset-8tc8g Successfully pulled image "quay.io/redhat-user-workloads/crt-redhat-acm-tenant/hive-operator/hive:hive-on-push-9v9lm-build-image-index" in 349ms (349ms including waiting). Image size: 1938808453 bytes. 114s Normal Created pod/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-imageset-8tc8g Created container: hiveutil 114s Normal Started pod/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-imageset-8tc8g Started container hiveutil 2m Normal SuccessfulCreate job/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-imageset Created pod: hive-7dda6e0a-a27e-4845-9609-3814aaddac46-imageset-8tc8g 111s Normal Completed job/hive-7dda6e0a-a27e-4845-9609-3814aaddac46-imageset Job completed 2026-02-02T08:40:04,384342710+00:00 Waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install 2026-02-02T08:41:04,557860693+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #1/180... 2026-02-02T08:42:04,748520800+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #2/180... 2026-02-02T08:43:04,930681347+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #3/180... 2026-02-02T08:44:05,124555755+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #4/180... 2026-02-02T08:45:05,314229305+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #5/180... 2026-02-02T08:46:05,516903449+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #6/180... 2026-02-02T08:47:05,712559610+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #7/180... 2026-02-02T08:48:05,910408654+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #8/180... 2026-02-02T08:49:06,103666898+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #9/180... 2026-02-02T08:50:06,284827359+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #10/180... 2026-02-02T08:51:06,476554003+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #11/180... 2026-02-02T08:52:06,660051600+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #12/180... 2026-02-02T08:53:06,858659718+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #13/180... 2026-02-02T08:54:07,051660866+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #14/180... 2026-02-02T08:55:07,236155760+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #15/180... 2026-02-02T08:56:07,429320492+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #16/180... 2026-02-02T08:57:07,625658787+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #17/180... 2026-02-02T08:58:07,801666430+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #18/180... 2026-02-02T08:59:07,983510278+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #19/180... 2026-02-02T09:00:08,184769755+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #20/180... 2026-02-02T09:01:08,363472870+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #21/180... 2026-02-02T09:02:08,544117988+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #22/180... 2026-02-02T09:03:08,744309017+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #23/180... 2026-02-02T09:04:08,916749283+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #24/180... 2026-02-02T09:05:09,101402962+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #25/180... 2026-02-02T09:06:09,283159697+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #26/180... 2026-02-02T09:07:09,459306260+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #27/180... 2026-02-02T09:08:09,645728439+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #28/180... 2026-02-02T09:09:09,835067061+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #29/180... 2026-02-02T09:10:10,022691690+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #30/180... 2026-02-02T09:11:10,200440513+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #31/180... 2026-02-02T09:12:10,401974720+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #32/180... 2026-02-02T09:13:10,600086291+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #33/180... 2026-02-02T09:14:10,802939783+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #34/180... 2026-02-02T09:15:10,981992868+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #35/180... 2026-02-02T09:16:11,185924250+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #36/180... 2026-02-02T09:17:11,379377198+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #37/180... 2026-02-02T09:18:11,580325872+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #38/180... 2026-02-02T09:19:11,773244442+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #39/180... 2026-02-02T09:20:11,959660742+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #40/180... 2026-02-02T09:21:12,156187267+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #41/180... 2026-02-02T09:22:12,364642183+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #42/180... 2026-02-02T09:23:12,584877004+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #43/180... 2026-02-02T09:24:12,773025289+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #44/180... 2026-02-02T09:25:12,976435799+00:00 Still waiting for the ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 to install. Status check #45/180... 2026-02-02T09:25:13,137068811+00:00 ClusterDeployment hive-7dda6e0a-a27e-4845-9609-3814aaddac46 was installed successfully 2026-02-02T09:25:13,272100132+00:00 Running post-install tests Defaulting to user installation because normal site-packages is not writeable Looking in links: /tmp/tmpeoid2ggz Requirement already satisfied: setuptools in /usr/lib/python3.9/site-packages (53.0.0) Requirement already satisfied: pip in /home/cloud-user/.local/lib/python3.9/site-packages (25.3) Requirement already satisfied: gitpython in /home/cloud-user/.local/lib/python3.9/site-packages (3.1.45) Requirement already satisfied: pyyaml in /usr/lib64/python3.9/site-packages (5.4.1) Requirement already satisfied: gitdb<5,>=4.0.1 in /home/cloud-user/.local/lib/python3.9/site-packages (from gitpython) (4.0.12) Requirement already satisfied: typing-extensions>=3.10.0.2 in /home/cloud-user/.local/lib/python3.9/site-packages (from gitpython) (4.14.1) Requirement already satisfied: smmap<6,>=3.0.1 in /home/cloud-user/.local/lib/python3.9/site-packages (from gitdb<5,>=4.0.1->gitpython) (5.0.2) [notice] A new release of pip is available: 25.3 -> 26.0 [notice] To update, run: python3 -m pip install --upgrade pip Using active branch HIVE-3067 since it corresponds to commit b8f7293 go test -mod=vendor -v -timeout 0 -count=1 ./test/e2e/postinstall/... === RUN TestScaleMachinePool time="2026-02-02T04:25:19-05:00" level=info msg="scaling pool to 1 replicas" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestScaleMachinePool time="2026-02-02T04:25:19-05:00" level=info msg="waiting for 1 machines (pool worker, infraID hive-7dda6e0a-a27e-48-zlhlh)" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestScaleMachinePool time="2026-02-02T04:25:19-05:00" level=info msg="Waiting for Machine" client=machine time="2026-02-02T04:25:19-05:00" level=info msg="found 2 machines for pool worker" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestScaleMachinePool time="2026-02-02T04:28:41-05:00" level=info msg="found 1 machines for pool worker" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestScaleMachinePool time="2026-02-02T04:28:41-05:00" level=info msg="waiting for 1 nodes (pool worker)" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestScaleMachinePool time="2026-02-02T04:28:41-05:00" level=info msg="Waiting for Nodes" client=node time="2026-02-02T04:28:42-05:00" level=info msg="scaling pool back to 3 replicas" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestScaleMachinePool time="2026-02-02T04:28:42-05:00" level=info msg="waiting for 3 machines (pool worker, infraID hive-7dda6e0a-a27e-48-zlhlh)" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestScaleMachinePool time="2026-02-02T04:28:42-05:00" level=info msg="Waiting for Machine" client=machine time="2026-02-02T04:28:42-05:00" level=info msg="found 1 machines for pool worker" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestScaleMachinePool time="2026-02-02T04:28:42-05:00" level=info msg="found 2 machines for pool worker" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestScaleMachinePool time="2026-02-02T04:28:42-05:00" level=info msg="found 3 machines for pool worker" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestScaleMachinePool time="2026-02-02T04:28:42-05:00" level=info msg="waiting for 3 nodes (pool worker)" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestScaleMachinePool time="2026-02-02T04:28:42-05:00" level=info msg="Waiting for Nodes" client=node --- PASS: TestScaleMachinePool (385.91s) === RUN TestNewMachinePool time="2026-02-02T04:31:45-05:00" level=info msg="Creating MachinePool" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-infra test=TestNewMachinePool infra_test.go:167: Waiting for 3 infra machines to be created time="2026-02-02T04:31:45-05:00" level=info msg="waiting for 3 machines (pool infra, infraID hive-7dda6e0a-a27e-48-zlhlh)" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-infra test=TestNewMachinePool time="2026-02-02T04:31:45-05:00" level=info msg="Waiting for Machine" client=machine time="2026-02-02T04:31:45-05:00" level=info msg="found 1 machines for pool infra" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-infra test=TestNewMachinePool time="2026-02-02T04:31:45-05:00" level=info msg="found 2 machines for pool infra" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-infra test=TestNewMachinePool time="2026-02-02T04:31:45-05:00" level=info msg="found 3 machines for pool infra" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-infra test=TestNewMachinePool time="2026-02-02T04:31:45-05:00" level=info msg="waiting for 3 nodes (pool infra)" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-infra test=TestNewMachinePool time="2026-02-02T04:31:45-05:00" level=info msg="Waiting for Nodes" client=node time="2026-02-02T04:35:11-05:00" level=info msg="removing pool" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-infra test=TestNewMachinePool time="2026-02-02T04:35:11-05:00" level=info msg="Waiting for MachineSet" client=machineset --- PASS: TestNewMachinePool (206.36s) === RUN TestAutoscalingMachinePool time="2026-02-02T04:35:11-05:00" level=info msg="switching pool from replicas to autoscaling" test=TestAutoscalingMachinePool time="2026-02-02T04:35:11-05:00" level=info msg="lowering autoscaler delay so scaling down happens faster" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestAutoscalingMachinePool infra_test.go:282: waiting for Hive to create cluster autoscaler infra_test.go:287: found cluster autoscaler time="2026-02-02T04:35:21-05:00" level=info msg="waiting for 12 machines (pool worker, infraID hive-7dda6e0a-a27e-48-zlhlh)" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestAutoscalingMachinePool time="2026-02-02T04:35:21-05:00" level=info msg="Waiting for Machine" client=machine time="2026-02-02T04:35:22-05:00" level=info msg="found 10 machines for pool worker" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestAutoscalingMachinePool time="2026-02-02T04:38:14-05:00" level=info msg="found 11 machines for pool worker" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestAutoscalingMachinePool time="2026-02-02T04:38:14-05:00" level=info msg="found 12 machines for pool worker" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestAutoscalingMachinePool time="2026-02-02T04:38:14-05:00" level=info msg="waiting for 12 nodes (pool worker)" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestAutoscalingMachinePool time="2026-02-02T04:38:14-05:00" level=info msg="Waiting for Nodes" client=node time="2026-02-02T04:41:30-05:00" level=info msg="deleting busybox deployment to relieve cpu pressure and scale down machines" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestAutoscalingMachinePool time="2026-02-02T04:41:30-05:00" level=info msg="waiting for 10 machines (pool worker, infraID hive-7dda6e0a-a27e-48-zlhlh)" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestAutoscalingMachinePool time="2026-02-02T04:41:30-05:00" level=info msg="Waiting for Machine" client=machine time="2026-02-02T04:41:30-05:00" level=info msg="found 12 machines for pool worker" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestAutoscalingMachinePool time="2026-02-02T04:43:32-05:00" level=info msg="found 11 machines for pool worker" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestAutoscalingMachinePool time="2026-02-02T04:43:33-05:00" level=info msg="found 10 machines for pool worker" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestAutoscalingMachinePool time="2026-02-02T04:43:33-05:00" level=info msg="waiting for 10 nodes (pool worker)" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestAutoscalingMachinePool time="2026-02-02T04:43:33-05:00" level=info msg="Waiting for Nodes" client=node time="2026-02-02T04:43:33-05:00" level=info msg="disabling autoscaling" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestAutoscalingMachinePool time="2026-02-02T04:43:33-05:00" level=info msg="waiting for 3 machines (pool worker, infraID hive-7dda6e0a-a27e-48-zlhlh)" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestAutoscalingMachinePool time="2026-02-02T04:43:33-05:00" level=info msg="Waiting for Machine" client=machine time="2026-02-02T04:43:33-05:00" level=info msg="found 10 machines for pool worker" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestAutoscalingMachinePool time="2026-02-02T04:45:08-05:00" level=info msg="found 9 machines for pool worker" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestAutoscalingMachinePool time="2026-02-02T04:45:29-05:00" level=info msg="found 8 machines for pool worker" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestAutoscalingMachinePool time="2026-02-02T04:45:58-05:00" level=info msg="found 7 machines for pool worker" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestAutoscalingMachinePool time="2026-02-02T04:46:44-05:00" level=info msg="found 6 machines for pool worker" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestAutoscalingMachinePool time="2026-02-02T04:47:06-05:00" level=info msg="found 5 machines for pool worker" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestAutoscalingMachinePool time="2026-02-02T04:47:46-05:00" level=info msg="found 4 machines for pool worker" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestAutoscalingMachinePool time="2026-02-02T04:47:50-05:00" level=info msg="found 3 machines for pool worker" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestAutoscalingMachinePool time="2026-02-02T04:47:50-05:00" level=info msg="waiting for 3 nodes (pool worker)" pool=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-worker test=TestAutoscalingMachinePool time="2026-02-02T04:47:50-05:00" level=info msg="Waiting for Nodes" client=node --- PASS: TestAutoscalingMachinePool (759.42s) PASS ok github.com/openshift/hive/test/e2e/postinstall/machinesets 1351.769s === RUN TestSyncsets Running Suite: Syncsets Suite ============================= Random Seed: 1770024319 Will run 4 of 4 specs E0202 04:25:19.833146 3077404 reflector.go:205] "Failed to watch" err="context canceled" logger="UnhandledError" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" type="*v1alpha1.ClusterSync" •••• Ran 4 of 4 Specs in 4.596 seconds SUCCESS! -- 4 Passed | 0 Failed | 0 Pending | 0 Skipped --- PASS: TestSyncsets (4.67s) PASS ok github.com/openshift/hive/test/e2e/postinstall/syncsets 4.713s 2026-02-02T09:47:53,123299053+00:00 ************* INSTALL JOB LOG ************* level=warning msg=Found override for release image (quay.io/openshift-release-dev/ocp-release:4.20.13-x86_64). Release Image Architecture is unknown level=info msg=Credentials loaded from default AWS environment variables level=warning msg=Length of cluster name "hive-7dda6e0a-a27e-4845-9609-3814aaddac46" is 41 which is greater than the max 21 allowed. The name will be truncated to "hive-7dda6e0a-a27e-48" level=info msg=Credentials loaded from the AWS config using "EnvConfigCredentials" provider level=info msg=Consuming Install Config from target directory level=info msg=Successfully populated MCS CA cert information: root-ca 2036-01-31T08:39:22Z 2026-02-02T08:39:22Z level=info msg=Successfully populated MCS TLS cert information: root-ca 2036-01-31T08:39:22Z 2026-02-02T08:39:22Z level=info msg=Adding clusters... level=info msg=Manifests created in: cluster-api, manifests and openshift level=warning msg=Found override for release image (quay.io/openshift-release-dev/ocp-release:4.20.13-x86_64). Please be warned, this is not advised level=info msg=Consuming OpenShift Install (Manifests) from target directory level=info msg=Consuming Worker Machines from target directory level=info msg=Consuming Openshift Manifests from target directory level=info msg=Consuming Common Manifests from target directory level=info msg=Consuming Master Machines from target directory level=info msg=Ignition-Configs created in: . and auth level=info msg=Consuming Worker Ignition Config from target directory level=info msg=Consuming Master Ignition Config from target directory level=info msg=Consuming Bootstrap Ignition Config from target directory level=info msg=Credentials loaded from default AWS environment variables level=info msg=Credentials loaded from the AWS config using "EnvConfigCredentials" provider level=info msg=Creating infrastructure resources... level=info msg=Reconciling IAM roles for control-plane and compute nodes level=info msg=Creating IAM role for master level=info msg=Creating IAM role for worker level=info msg=Started local control plane with envtest level=info msg=Stored kubeconfig for envtest in: /output/.clusterapi_output/envtest.kubeconfig level=info msg=Running process: Cluster API with args [-v=2 --diagnostics-address=0 --health-addr=127.0.0.1:46263 --webhook-port=33151 --webhook-cert-dir=/tmp/envtest-serving-certs-2212988566 --kubeconfig=/output/.clusterapi_output/envtest.kubeconfig] level=info msg=Running process: aws infrastructure provider with args [-v=4 --diagnostics-address=0 --health-addr=127.0.0.1:35905 --webhook-port=44639 --webhook-cert-dir=/tmp/envtest-serving-certs-4277582489 --feature-gates=BootstrapFormatIgnition=true,ExternalResourceGC=true,TagUnmanagedNetworkResources=false,EKS=false --kubeconfig=/output/.clusterapi_output/envtest.kubeconfig] level=info msg=Creating infra manifests... level=info msg=Created manifest *v1.Namespace, namespace= name=openshift-cluster-api-guests level=info msg=Created manifest *v1beta2.AWSClusterControllerIdentity, namespace= name=default level=info msg=Created manifest *v1beta1.Cluster, namespace=openshift-cluster-api-guests name=hive-7dda6e0a-a27e-48-zlhlh level=info msg=Created manifest *v1beta2.AWSCluster, namespace=openshift-cluster-api-guests name=hive-7dda6e0a-a27e-48-zlhlh level=info msg=Done creating infra manifests level=info msg=Creating kubeconfig entry for capi cluster hive-7dda6e0a-a27e-48-zlhlh level=info msg=Waiting up to 15m0s (until 8:54AM UTC) for network infrastructure to become ready... level=info msg=Network infrastructure is ready level=info msg=Creating Route53 records for control plane load balancer level=info msg=Created private Hosted Zone level=info msg=Created manifest *v1beta2.AWSMachine, namespace=openshift-cluster-api-guests name=hive-7dda6e0a-a27e-48-zlhlh-bootstrap level=info msg=Created manifest *v1beta2.AWSMachine, namespace=openshift-cluster-api-guests name=hive-7dda6e0a-a27e-48-zlhlh-master-0 level=info msg=Created manifest *v1beta2.AWSMachine, namespace=openshift-cluster-api-guests name=hive-7dda6e0a-a27e-48-zlhlh-master-1 level=info msg=Created manifest *v1beta2.AWSMachine, namespace=openshift-cluster-api-guests name=hive-7dda6e0a-a27e-48-zlhlh-master-2 level=info msg=Created manifest *v1beta1.Machine, namespace=openshift-cluster-api-guests name=hive-7dda6e0a-a27e-48-zlhlh-bootstrap level=info msg=Created manifest *v1beta1.Machine, namespace=openshift-cluster-api-guests name=hive-7dda6e0a-a27e-48-zlhlh-master-0 level=info msg=Created manifest *v1beta1.Machine, namespace=openshift-cluster-api-guests name=hive-7dda6e0a-a27e-48-zlhlh-master-1 level=info msg=Created manifest *v1beta1.Machine, namespace=openshift-cluster-api-guests name=hive-7dda6e0a-a27e-48-zlhlh-master-2 level=info msg=Created manifest *v1.Secret, namespace=openshift-cluster-api-guests name=hive-7dda6e0a-a27e-48-zlhlh-bootstrap level=info msg=Created manifest *v1.Secret, namespace=openshift-cluster-api-guests name=hive-7dda6e0a-a27e-48-zlhlh-master level=info msg=Created manifest *v1.Secret, namespace=openshift-cluster-api-guests name=hive-7dda6e0a-a27e-48-zlhlh-worker level=info msg=Waiting up to 15m0s (until 9:02AM UTC) for machines [hive-7dda6e0a-a27e-48-zlhlh-bootstrap hive-7dda6e0a-a27e-48-zlhlh-master-0 hive-7dda6e0a-a27e-48-zlhlh-master-1 hive-7dda6e0a-a27e-48-zlhlh-master-2] to provision... level=info msg=Control-plane machines are ready level=info msg=Cluster API resources have been created. Waiting for cluster to become ready... level=info msg=Consuming Cluster API Machine Manifests from target directory level=info msg=Consuming Cluster API Manifests from target directory level=info msg=Waiting up to 20m0s (until 9:08AM UTC) for the Kubernetes API at https://api.hive-7dda6e0a-a27e-4845-9609-3814aaddac46.qe.devcluster.openshift.com:6443... level=info msg=API v1.33.6 up level=info msg=Waiting up to 45m0s (until 9:36AM UTC) for bootstrapping to complete... level=info msg=Waiting for the bootstrap etcd member to be removed... level=info msg=Bootstrap etcd member has been removed level=info msg=Destroying the bootstrap resources... level=info msg=Waiting up to 5m0s for bootstrap machine deletion openshift-cluster-api-guests/hive-7dda6e0a-a27e-48-zlhlh-bootstrap... level=info msg=Shutting down local Cluster API controllers... level=info msg=Stopped controller: Cluster API level=info msg=Stopped controller: aws infrastructure provider level=info msg=Shutting down local Cluster API control plane... level=info msg=Local Cluster API system has completed operations level=info msg=Finished destroying bootstrap resources level=info msg=Waiting up to 40m0s (until 9:48AM UTC) for the cluster at https://api.hive-7dda6e0a-a27e-4845-9609-3814aaddac46.qe.devcluster.openshift.com:6443 to initialize... level=info msg=Waiting up to 30m0s (until 9:52AM UTC) to ensure each cluster operator has finished progressing... level=info msg=All cluster operators have completed progressing level=info msg=Checking to see if there is a route at openshift-console/console... level=info msg=Install complete! level=info msg=To access the cluster as the system:admin user when using 'oc', run level=info msg= export KUBECONFIG=/output/auth/kubeconfig level=info msg=Access the OpenShift web-console here: https://console-openshift-console.apps.hive-7dda6e0a-a27e-4845-9609-3814aaddac46.qe.devcluster.openshift.com REDACTED LINE OF OUTPUT level=info msg=Time elapsed: 43m36s 2026-02-02T09:47:53,266405233+00:00 Running destroy test Defaulting to user installation because normal site-packages is not writeable Looking in links: /tmp/tmpvccewjgh Requirement already satisfied: setuptools in /usr/lib/python3.9/site-packages (53.0.0) Requirement already satisfied: pip in /home/cloud-user/.local/lib/python3.9/site-packages (25.3) Requirement already satisfied: gitpython in /home/cloud-user/.local/lib/python3.9/site-packages (3.1.45) Requirement already satisfied: pyyaml in /usr/lib64/python3.9/site-packages (5.4.1) Requirement already satisfied: gitdb<5,>=4.0.1 in /home/cloud-user/.local/lib/python3.9/site-packages (from gitpython) (4.0.12) Requirement already satisfied: typing-extensions>=3.10.0.2 in /home/cloud-user/.local/lib/python3.9/site-packages (from gitpython) (4.14.1) Requirement already satisfied: smmap<6,>=3.0.1 in /home/cloud-user/.local/lib/python3.9/site-packages (from gitdb<5,>=4.0.1->gitpython) (5.0.2) [notice] A new release of pip is available: 25.3 -> 26.0 [notice] To update, run: python3 -m pip install --upgrade pip Using active branch HIVE-3067 since it corresponds to commit b8f7293 go test -mod=vendor -v -timeout 0 -count=1 ./test/e2e/destroycluster/... === RUN TestDestroyCluster time="2026-02-02T04:47:58-05:00" level=info msg="Starting destroy cluster test" test=TestDestroyCluster time="2026-02-02T04:47:58-05:00" level=info msg="Deleting cluster deployment" cluster=hive-e2e1/hive-7dda6e0a-a27e-4845-9609-3814aaddac46 test=TestDestroyCluster time="2026-02-02T04:47:58-05:00" level=info msg="Waiting for cluster deployment to go away" cluster=hive-e2e1/hive-7dda6e0a-a27e-4845-9609-3814aaddac46 test=TestDestroyCluster time="2026-02-02T04:47:58-05:00" level=info msg="Deprovision job created" cluster=hive-e2e1/hive-7dda6e0a-a27e-4845-9609-3814aaddac46 job=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-uninstall test=TestDestroyCluster time="2026-02-02T04:48:01-05:00" level=info msg="Saving logs to /home/cloud-user/mihuang/hive/mihuang/artifacts/hive_uninstall_job.log" cluster=hive-e2e1/hive-7dda6e0a-a27e-4845-9609-3814aaddac46 job=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-uninstall test=TestDestroyCluster time="2026-02-02T04:52:50-05:00" level=info msg="Deprovision job has finished" cluster=hive-e2e1/hive-7dda6e0a-a27e-4845-9609-3814aaddac46 job=hive-7dda6e0a-a27e-4845-9609-3814aaddac46-uninstall test=TestDestroyCluster time="2026-02-02T04:52:50-05:00" level=info msg="Cluster deployment has been removed" cluster=hive-e2e1/hive-7dda6e0a-a27e-4845-9609-3814aaddac46 test=TestDestroyCluster time="2026-02-02T04:52:50-05:00" level=info msg="No DNS Zones related to the cluster exist" cluster=hive-e2e1/hive-7dda6e0a-a27e-4845-9609-3814aaddac46 test=TestDestroyCluster --- PASS: TestDestroyCluster (292.42s) PASS ok github.com/openshift/hive/test/e2e/destroycluster 292.467s 2026-02-02T09:52:50,884732205+00:00 Saving hive logs Found 2 pods, using pod/hiveadmission-78b7cd97bb-zknwx 2026-02-02T09:52:56,290316182+00:00 Uninstalling hive and validating cleanup Defaulting to user installation because normal site-packages is not writeable Looking in links: /tmp/tmpx7r9hu65 Requirement already satisfied: setuptools in /usr/lib/python3.9/site-packages (53.0.0) Requirement already satisfied: pip in /home/cloud-user/.local/lib/python3.9/site-packages (25.3) Requirement already satisfied: gitpython in /home/cloud-user/.local/lib/python3.9/site-packages (3.1.45) Requirement already satisfied: pyyaml in /usr/lib64/python3.9/site-packages (5.4.1) Requirement already satisfied: gitdb<5,>=4.0.1 in /home/cloud-user/.local/lib/python3.9/site-packages (from gitpython) (4.0.12) Requirement already satisfied: typing-extensions>=3.10.0.2 in /home/cloud-user/.local/lib/python3.9/site-packages (from gitpython) (4.14.1) Requirement already satisfied: smmap<6,>=3.0.1 in /home/cloud-user/.local/lib/python3.9/site-packages (from gitdb<5,>=4.0.1->gitpython) (5.0.2) [notice] A new release of pip is available: 25.3 -> 26.0 [notice] To update, run: python3 -m pip install --upgrade pip Using active branch HIVE-3067 since it corresponds to commit b8f7293 go test -mod=vendor -v -timeout 0 -count=1 ./test/e2e/uninstallhive/... === RUN TestUninstallHive time="2026-02-02T04:53:01-05:00" level=info msg="deleting HiveConfig 'hive'" test=TestUninstallHive time="2026-02-02T04:53:01-05:00" level=info msg="resource successfully deleted" gvr="hive.openshift.io/v1, Resource=hiveconfigs" name=hive namespace= test=TestUninstallHive === RUN TestUninstallHive/hive-controllers_deleted time="2026-02-02T04:53:01-05:00" level=info msg="resource successfully deleted" gvr="apps/v1, Resource=deployments" name=hive-controllers namespace=hive-e2e-two test=TestUninstallHive time="2026-02-02T04:53:01-05:00" level=info msg="resource still exists, sleeping..." gvr="/v1, Resource=services" name=hive-controllers namespace=hive-e2e-two test=TestUninstallHive time="2026-02-02T04:53:01-05:00" level=info msg="resource successfully deleted" gvr="/v1, Resource=services" name=hive-controllers namespace=hive-e2e-two test=TestUninstallHive time="2026-02-02T04:53:01-05:00" level=info msg="resource successfully deleted" gvr="/v1, Resource=serviceaccounts" name=hive-controllers namespace=hive-e2e-two test=TestUninstallHive time="2026-02-02T04:53:01-05:00" level=info msg="resource successfully deleted" gvr="/v1, Resource=serviceaccounts" name=hive-frontend namespace=hive-e2e-two test=TestUninstallHive time="2026-02-02T04:53:01-05:00" level=info msg="resource successfully deleted" gvr="rbac.authorization.k8s.io/v1, Resource=clusterroles" name=hive-controllers namespace= test=TestUninstallHive time="2026-02-02T04:53:01-05:00" level=info msg="resource successfully deleted" gvr="rbac.authorization.k8s.io/v1, Resource=clusterroles" name=hive-frontend namespace= test=TestUninstallHive === RUN TestUninstallHive/hiveadmission_deleted time="2026-02-02T04:53:01-05:00" level=info msg="resource successfully deleted" gvr="apps/v1, Resource=deployments" name=hiveadmission namespace=hive-e2e-two test=TestUninstallHive time="2026-02-02T04:53:01-05:00" level=info msg="resource successfully deleted" gvr="/v1, Resource=services" name=hiveadmission namespace=hive-e2e-two test=TestUninstallHive time="2026-02-02T04:53:01-05:00" level=info msg="resource successfully deleted" gvr="/v1, Resource=serviceaccounts" name=hiveadmission namespace=hive-e2e-two test=TestUninstallHive time="2026-02-02T04:53:01-05:00" level=info msg="resource successfully deleted" gvr="rbac.authorization.k8s.io/v1, Resource=system:openshift:hive:hiveadmission" name=hive-controllers namespace= test=TestUninstallHive time="2026-02-02T04:53:01-05:00" level=info msg="resource successfully deleted" gvr="apiregistration.k8s.io/v1, Resource=apiservices" name=v1.admission.hive.openshift.io namespace= test=TestUninstallHive --- PASS: TestUninstallHive (0.76s) --- PASS: TestUninstallHive/hive-controllers_deleted (0.62s) --- PASS: TestUninstallHive/hiveadmission_deleted (0.06s) PASS ok github.com/openshift/hive/test/e2e/uninstallhive 0.807s 2026-02-02T09:53:01,897383992+00:00 [1]+ Terminated /usr/bin/bash -c "sleep $(($timeout_minutes*60)) && echo 'Timed out!' && kill -n 2 $$" !EXIT TRAP! 2026-02-02T09:53:03,401889468+00:00 Saving hive logs before cleanup error: error from server (NotFound): deployments.apps "hive-controllers" not found in namespace "hive-e2e-two" No resources found in hive-e2e-two namespace. error: error from server (NotFound): deployments.apps "hiveadmission" not found in namespace "hive-e2e-two" No resources found in hive-e2e-two namespace. error: error from server (NotFound): statefulsets.apps "hive-clustersync" not found in namespace "hive-e2e-two" No resources found in hive-e2e-two namespace. error: error from server (NotFound): statefulsets.apps "hive-machinepool" not found in namespace "hive-e2e-two" No resources found in hive-e2e-two namespace. No resources found in hive-e2e-two namespace. Error from server (NotFound): hiveconfigs.hive.openshift.io "hive" not found Error from server (NotFound): apiservices.apiregistration.k8s.io "v1.admission.hive.openshift.io" not found 2026-02-02T09:53:06,583312836+00:00 2026-02-02T09:53:06,587554814+00:00 Error from server (NotFound): clusterdeployments.hive.openshift.io "hive-7dda6e0a-a27e-4845-9609-3814aaddac46" not found