% Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 3831 100 3831 0 0 18096 0 --:--:-- --:--:-- --:--:-- 18156 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 9315 100 9315 0 0 44297 0 --:--:-- --:--:-- --:--:-- 44146 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 2607 100 2607 0 0 15575 0 --:--:-- --:--:-- --:--:-- 15610 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 1953 100 1953 0 0 10817 0 --:--:-- --:--:-- --:--:-- 10850 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 350 100 350 0 0 2373 0 --:--:-- --:--:-- --:--:-- 2380 Installing kuttl Try 0... curl -sLo /tmp/jaeger-tests/hack/install/../../bin/kubectl-kuttl https://github.com/kudobuilder/kuttl/releases/download/v0.15.0/kubectl-kuttl_0.15.0_linux_x86_64 KUBECONFIG file is: /tmp/kubeconfig-2856073267 for suite in streaming sidecar ui miscellaneous elasticsearch examples generate upgrade; do \ make run-e2e-tests-$suite ; \ done make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh streaming false true + '[' 3 -ne 3 ']' + test_suite_name=streaming + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/streaming.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-streaming make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ SKIP_ES_EXTERNAL=true \ ./tests/e2e/streaming/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.17.0-0.nightly-2024-09-29-173421 True False 10m Cluster version is 4.17.0-0.nightly-2024-09-29-173421' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.17.0-0.nightly-2024-09-29-173421 True False 10m Cluster version is 4.17.0-0.nightly-2024-09-29-173421' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/streaming/render.sh ++ export SUITE_DIR=./tests/e2e/streaming ++ SUITE_DIR=./tests/e2e/streaming ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/streaming ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + '[' false = true ']' + start_test streaming-simple + '[' 1 -ne 1 ']' + test_name=streaming-simple + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-simple' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-simple\e[0m' Rendering files for test streaming-simple + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build + '[' _build '!=' _build ']' + mkdir -p streaming-simple + cd streaming-simple + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/streaming-jaeger-assert.yaml.template -o ./04-assert.yaml + render_smoke_test simple-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=simple-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + export JAEGER_NAME=simple-streaming + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-tls + '[' 1 -ne 1 ']' + test_name=streaming-with-tls + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-tls' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-tls\e[0m' Rendering files for test streaming-with-tls + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-simple + '[' streaming-simple '!=' _build ']' + cd .. + mkdir -p streaming-with-tls + cd streaming-with-tls + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + render_smoke_test tls-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=tls-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + export JAEGER_NAME=tls-streaming + JAEGER_NAME=tls-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-autoprovisioning-autoscale + '[' 1 -ne 1 ']' + test_name=streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-autoprovisioning-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-autoprovisioning-autoscale\e[0m' Rendering files for test streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-with-tls + '[' streaming-with-tls '!=' _build ']' + cd .. + mkdir -p streaming-with-autoprovisioning-autoscale + cd streaming-with-autoprovisioning-autoscale + '[' true = true ']' + rm ./00-install.yaml ./00-assert.yaml + render_install_elasticsearch upstream 01 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=01 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./01-assert.yaml + jaeger_name=auto-provisioned + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="20Mi"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="500m"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.autoscale=true ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.minReplicas=1 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.maxReplicas=2 ./02-install.yaml + render_assert_kafka true auto-provisioned 03 + '[' 3 -ne 3 ']' + autoprovisioned=true + cluster_name=auto-provisioned + test_step=03 + '[' true = true ']' + is_kafka_minimal_enabled + namespaces=(observability openshift-operators openshift-distributed-tracing) + for i in "${namespaces[@]}" ++ kubectl get pods -n observability -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-operators -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-distributed-tracing -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled=true + '[' true == true ']' + return 0 + replicas=1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./03-assert.yaml ++ expr 03 + 1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./04-assert.yaml ++ expr 03 + 2 + CLUSTER_NAME=auto-provisioned + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./05-assert.yaml + version_lt 1.30 1.23 ++ echo 1.30 1.23 ++ tr ' ' '\n' ++ sort -rV ++ head -n 1 + test 1.30 '!=' 1.30 + rm ./08-assert.yaml + skip_test streaming-with-tls 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=streaming-with-tls + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-with-autoprovisioning-autoscale + '[' streaming-with-autoprovisioning-autoscale '!=' _build ']' + cd .. + rm -rf streaming-with-tls + warning 'streaming-with-tls: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: streaming-with-tls: This test is flaky in Prow CI\e[0m' WAR: streaming-with-tls: This test is flaky in Prow CI + skip_test streaming-simple 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=streaming-simple + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build + '[' _build '!=' _build ']' + rm -rf streaming-simple + warning 'streaming-simple: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: streaming-simple: This test is flaky in Prow CI\e[0m' WAR: streaming-simple: This test is flaky in Prow CI make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running streaming E2E tests' Running streaming E2E tests + cd tests/e2e/streaming/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2856073267 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 2 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/streaming-with-autoprovisioning-autoscale === PAUSE kuttl/harness/streaming-with-autoprovisioning-autoscale === CONT kuttl/harness/artifacts logger.go:42: 17:48:04 | artifacts | Creating namespace: kuttl-test-causal-cheetah logger.go:42: 17:48:04 | artifacts | artifacts events from ns kuttl-test-causal-cheetah: logger.go:42: 17:48:04 | artifacts | Deleting namespace: kuttl-test-causal-cheetah === CONT kuttl/harness/streaming-with-autoprovisioning-autoscale logger.go:42: 17:48:10 | streaming-with-autoprovisioning-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 17:48:10 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_0.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 17:48:10 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_1.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 17:48:10 | streaming-with-autoprovisioning-autoscale | Creating namespace: kuttl-test-bold-jawfish logger.go:42: 17:48:10 | streaming-with-autoprovisioning-autoscale/1-install | starting test step 1-install logger.go:42: 17:48:10 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc create sa deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 17:48:10 | streaming-with-autoprovisioning-autoscale/1-install | serviceaccount/deploy-elasticsearch created logger.go:42: 17:48:10 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc adm policy add-scc-to-user privileged -z deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 17:48:11 | streaming-with-autoprovisioning-autoscale/1-install | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:privileged added: "deploy-elasticsearch" logger.go:42: 17:48:11 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 6] logger.go:42: 17:48:17 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_0.yml -n $NAMESPACE] logger.go:42: 17:48:17 | streaming-with-autoprovisioning-autoscale/1-install | statefulset.apps/elasticsearch created logger.go:42: 17:48:17 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 3] logger.go:42: 17:48:20 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_1.yml -n $NAMESPACE] logger.go:42: 17:48:20 | streaming-with-autoprovisioning-autoscale/1-install | service/elasticsearch created logger.go:42: 17:48:38 | streaming-with-autoprovisioning-autoscale/1-install | test step completed 1-install logger.go:42: 17:48:38 | streaming-with-autoprovisioning-autoscale/2-install | starting test step 2-install logger.go:42: 17:48:38 | streaming-with-autoprovisioning-autoscale/2-install | Jaeger:kuttl-test-bold-jawfish/auto-provisioned created logger.go:42: 17:48:38 | streaming-with-autoprovisioning-autoscale/2-install | test step completed 2-install logger.go:42: 17:48:38 | streaming-with-autoprovisioning-autoscale/3- | starting test step 3- logger.go:42: 17:49:23 | streaming-with-autoprovisioning-autoscale/3- | test step completed 3- logger.go:42: 17:49:23 | streaming-with-autoprovisioning-autoscale/4- | starting test step 4- logger.go:42: 17:49:53 | streaming-with-autoprovisioning-autoscale/4- | test step completed 4- logger.go:42: 17:49:53 | streaming-with-autoprovisioning-autoscale/5- | starting test step 5- logger.go:42: 17:50:16 | streaming-with-autoprovisioning-autoscale/5- | test step completed 5- logger.go:42: 17:50:16 | streaming-with-autoprovisioning-autoscale/6- | starting test step 6- logger.go:42: 17:50:32 | streaming-with-autoprovisioning-autoscale/6- | test step completed 6- logger.go:42: 17:50:32 | streaming-with-autoprovisioning-autoscale/7- | starting test step 7- logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale/7- | test step completed 7- logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | streaming-with-autoprovisioning-autoscale events from ns kuttl-test-bold-jawfish: logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:48:17 +0000 UTC Normal Pod elasticsearch-0 Binding Scheduled Successfully assigned kuttl-test-bold-jawfish/elasticsearch-0 to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:48:17 +0000 UTC Normal StatefulSet.apps elasticsearch SuccessfulCreate create Pod elasticsearch-0 in StatefulSet elasticsearch successful statefulset-controller logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:48:18 +0000 UTC Normal Pod elasticsearch-0 AddedInterface Add eth0 [10.128.2.14/23] from ovn-kubernetes multus logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:48:18 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulling Pulling image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:48:26 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulled Successfully pulled image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" in 7.908s (7.908s including waiting). Image size: 758467647 bytes. kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:48:26 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:48:26 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:48:33 +0000 UTC Warning Pod elasticsearch-0.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Get "http://10.128.2.14:9200/": dial tcp 10.128.2.14:9200: connect: connection refused kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:48:45 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:48:46 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-zookeeper NoPods No matching pods found controllermanager logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:48:46 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-bold-jawfish/data-auto-provisioned-zookeeper-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-58df948849-24kf7_d8dc002d-38cb-4fd1-83cb-f837184821df logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:48:46 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:48:48 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ProvisioningSucceeded Successfully provisioned volume pvc-56be4f1f-f885-41b4-a4d7-999fd72951d8 ebs.csi.aws.com_aws-ebs-csi-driver-controller-58df948849-24kf7_d8dc002d-38cb-4fd1-83cb-f837184821df logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:48:49 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 Binding Scheduled Successfully assigned kuttl-test-bold-jawfish/auto-provisioned-zookeeper-0 to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:48:51 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-56be4f1f-f885-41b4-a4d7-999fd72951d8" attachdetach-controller logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:48:52 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 AddedInterface Add eth0 [10.131.0.26/23] from ovn-kubernetes multus logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:48:52 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulling Pulling image "registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:917efd492885c2071fbf06372348f812ba600b6788052e2700bd730c7f4b794e" kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:49:03 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulled Successfully pulled image "registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:917efd492885c2071fbf06372348f812ba600b6788052e2700bd730c7f4b794e" in 11.077s (11.077s including waiting). Image size: 666780849 bytes. kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:49:03 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Created Created container zookeeper kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:49:03 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Started Started container zookeeper kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:49:23 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:49:24 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-kafka NoPods No matching pods found controllermanager logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:49:24 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:49:24 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-bold-jawfish/data-0-auto-provisioned-kafka-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-58df948849-24kf7_d8dc002d-38cb-4fd1-83cb-f837184821df logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:49:26 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ProvisioningSucceeded Successfully provisioned volume pvc-ce6e109d-3b1a-4c9a-af1c-837dd66f685c ebs.csi.aws.com_aws-ebs-csi-driver-controller-58df948849-24kf7_d8dc002d-38cb-4fd1-83cb-f837184821df logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:49:27 +0000 UTC Normal Pod auto-provisioned-kafka-0 Binding Scheduled Successfully assigned kuttl-test-bold-jawfish/auto-provisioned-kafka-0 to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:49:30 +0000 UTC Normal Pod auto-provisioned-kafka-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-ce6e109d-3b1a-4c9a-af1c-837dd66f685c" attachdetach-controller logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:49:33 +0000 UTC Normal Pod auto-provisioned-kafka-0 AddedInterface Add eth0 [10.131.0.27/23] from ovn-kubernetes multus logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:49:33 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Pulled Container image "registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:917efd492885c2071fbf06372348f812ba600b6788052e2700bd730c7f4b794e" already present on machine kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:49:33 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Created Created container kafka kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:49:33 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Started Started container kafka kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:49:55 +0000 UTC Normal Pod auto-provisioned-entity-operator-84946db49f-l59sz Binding Scheduled Successfully assigned kuttl-test-bold-jawfish/auto-provisioned-entity-operator-84946db49f-l59sz to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:49:55 +0000 UTC Normal ReplicaSet.apps auto-provisioned-entity-operator-84946db49f SuccessfulCreate Created pod: auto-provisioned-entity-operator-84946db49f-l59sz replicaset-controller logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:49:55 +0000 UTC Normal Deployment.apps auto-provisioned-entity-operator ScalingReplicaSet Scaled up replica set auto-provisioned-entity-operator-84946db49f to 1 deployment-controller logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:49:56 +0000 UTC Normal Pod auto-provisioned-entity-operator-84946db49f-l59sz AddedInterface Add eth0 [10.129.2.20/23] from ovn-kubernetes multus logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:49:56 +0000 UTC Normal Pod auto-provisioned-entity-operator-84946db49f-l59sz.spec.containers{topic-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:d1c577cce87e2063b483db3a70bac25c9b5c7fa1d68fbdc7774d2a71c2f8e8a1" already present on machine kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:49:56 +0000 UTC Normal Pod auto-provisioned-entity-operator-84946db49f-l59sz.spec.containers{topic-operator} Created Created container topic-operator kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:49:56 +0000 UTC Normal Pod auto-provisioned-entity-operator-84946db49f-l59sz.spec.containers{topic-operator} Started Started container topic-operator kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:49:56 +0000 UTC Normal Pod auto-provisioned-entity-operator-84946db49f-l59sz.spec.containers{user-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:d1c577cce87e2063b483db3a70bac25c9b5c7fa1d68fbdc7774d2a71c2f8e8a1" already present on machine kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:49:56 +0000 UTC Normal Pod auto-provisioned-entity-operator-84946db49f-l59sz.spec.containers{user-operator} Created Created container user-operator kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:49:56 +0000 UTC Normal Pod auto-provisioned-entity-operator-84946db49f-l59sz.spec.containers{user-operator} Started Started container user-operator kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:19 +0000 UTC Normal Pod auto-provisioned-collector-7bcc7d5f9b-pbpzg Binding Scheduled Successfully assigned kuttl-test-bold-jawfish/auto-provisioned-collector-7bcc7d5f9b-pbpzg to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:19 +0000 UTC Normal ReplicaSet.apps auto-provisioned-collector-7bcc7d5f9b SuccessfulCreate Created pod: auto-provisioned-collector-7bcc7d5f9b-pbpzg replicaset-controller logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:19 +0000 UTC Normal Deployment.apps auto-provisioned-collector ScalingReplicaSet Scaled up replica set auto-provisioned-collector-7bcc7d5f9b to 1 deployment-controller logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:19 +0000 UTC Normal Pod auto-provisioned-ingester-5996d4f447-xp6gl Binding Scheduled Successfully assigned kuttl-test-bold-jawfish/auto-provisioned-ingester-5996d4f447-xp6gl to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:19 +0000 UTC Normal ReplicaSet.apps auto-provisioned-ingester-5996d4f447 SuccessfulCreate Created pod: auto-provisioned-ingester-5996d4f447-xp6gl replicaset-controller logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:19 +0000 UTC Normal Deployment.apps auto-provisioned-ingester ScalingReplicaSet Scaled up replica set auto-provisioned-ingester-5996d4f447 to 1 deployment-controller logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:19 +0000 UTC Normal Pod auto-provisioned-query-67b76955ff-qt5nj Binding Scheduled Successfully assigned kuttl-test-bold-jawfish/auto-provisioned-query-67b76955ff-qt5nj to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:19 +0000 UTC Normal ReplicaSet.apps auto-provisioned-query-67b76955ff SuccessfulCreate Created pod: auto-provisioned-query-67b76955ff-qt5nj replicaset-controller logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:19 +0000 UTC Normal Deployment.apps auto-provisioned-query ScalingReplicaSet Scaled up replica set auto-provisioned-query-67b76955ff to 1 deployment-controller logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:20 +0000 UTC Normal Pod auto-provisioned-collector-7bcc7d5f9b-pbpzg AddedInterface Add eth0 [10.128.2.15/23] from ovn-kubernetes multus logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:20 +0000 UTC Normal Pod auto-provisioned-collector-7bcc7d5f9b-pbpzg.spec.containers{jaeger-collector} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:20 +0000 UTC Normal Pod auto-provisioned-ingester-5996d4f447-xp6gl AddedInterface Add eth0 [10.129.2.21/23] from ovn-kubernetes multus logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:20 +0000 UTC Normal Pod auto-provisioned-ingester-5996d4f447-xp6gl.spec.containers{jaeger-ingester} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:7f435686c86e24ca85e6ebfa2f3eb5a456a81a8c2d4df56e3f7b68fff46eaf85" kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:20 +0000 UTC Normal Pod auto-provisioned-query-67b76955ff-qt5nj AddedInterface Add eth0 [10.131.0.28/23] from ovn-kubernetes multus logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:20 +0000 UTC Normal Pod auto-provisioned-query-67b76955ff-qt5nj.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:23 +0000 UTC Normal Pod auto-provisioned-collector-7bcc7d5f9b-pbpzg.spec.containers{jaeger-collector} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" in 3.794s (3.794s including waiting). Image size: 142315153 bytes. kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:23 +0000 UTC Normal Pod auto-provisioned-collector-7bcc7d5f9b-pbpzg.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:23 +0000 UTC Normal Pod auto-provisioned-collector-7bcc7d5f9b-pbpzg.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:24 +0000 UTC Normal Pod auto-provisioned-query-67b76955ff-qt5nj.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" in 4.042s (4.042s including waiting). Image size: 192762590 bytes. kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:24 +0000 UTC Normal Pod auto-provisioned-query-67b76955ff-qt5nj.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:24 +0000 UTC Normal Pod auto-provisioned-query-67b76955ff-qt5nj.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:24 +0000 UTC Normal Pod auto-provisioned-query-67b76955ff-qt5nj.spec.containers{oauth-proxy} Pulling Pulling image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:26 +0000 UTC Normal Pod auto-provisioned-ingester-5996d4f447-xp6gl.spec.containers{jaeger-ingester} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:7f435686c86e24ca85e6ebfa2f3eb5a456a81a8c2d4df56e3f7b68fff46eaf85" in 6.188s (6.189s including waiting). Image size: 139591266 bytes. kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:26 +0000 UTC Normal Pod auto-provisioned-ingester-5996d4f447-xp6gl.spec.containers{jaeger-ingester} Created Created container jaeger-ingester kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:26 +0000 UTC Normal Pod auto-provisioned-ingester-5996d4f447-xp6gl.spec.containers{jaeger-ingester} Started Started container jaeger-ingester kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:27 +0000 UTC Warning Pod auto-provisioned-ingester-5996d4f447-xp6gl.spec.containers{jaeger-ingester} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:28 +0000 UTC Normal Pod auto-provisioned-query-67b76955ff-qt5nj.spec.containers{oauth-proxy} Pulled Successfully pulled image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" in 3.744s (3.744s including waiting). Image size: 339954870 bytes. kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:28 +0000 UTC Normal Pod auto-provisioned-query-67b76955ff-qt5nj.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:28 +0000 UTC Normal Pod auto-provisioned-query-67b76955ff-qt5nj.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:28 +0000 UTC Normal Pod auto-provisioned-query-67b76955ff-qt5nj.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:30 +0000 UTC Normal Pod auto-provisioned-query-67b76955ff-qt5nj.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" in 1.798s (1.798s including waiting). Image size: 115523131 bytes. kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:30 +0000 UTC Normal Pod auto-provisioned-query-67b76955ff-qt5nj.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | 2024-09-30 17:50:30 +0000 UTC Normal Pod auto-provisioned-query-67b76955ff-qt5nj.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 17:50:33 | streaming-with-autoprovisioning-autoscale | Deleting namespace: kuttl-test-bold-jawfish === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (167.18s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.91s) --- PASS: kuttl/harness/streaming-with-autoprovisioning-autoscale (161.13s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name streaming --report --output /logs/artifacts/streaming.xml ./artifacts/kuttl-report.xml time="2024-09-30T17:50:53Z" level=debug msg="Setting a new name for the test suites" time="2024-09-30T17:50:53Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-09-30T17:50:53Z" level=debug msg="normalizing test case names" time="2024-09-30T17:50:53Z" level=debug msg="streaming/artifacts -> streaming_artifacts" time="2024-09-30T17:50:53Z" level=debug msg="streaming/streaming-with-autoprovisioning-autoscale -> streaming_streaming_with_autoprovisioning_autoscale" +-----------------------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------------------+--------+ | streaming_artifacts | passed | | streaming_streaming_with_autoprovisioning_autoscale | passed | +-----------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh sidecar false true + '[' 3 -ne 3 ']' + test_suite_name=sidecar + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/sidecar.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-sidecar make[2]: Entering directory '/tmp/jaeger-tests' ./tests/e2e/sidecar/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.17.0-0.nightly-2024-09-29-173421 True False 13m Cluster version is 4.17.0-0.nightly-2024-09-29-173421' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.17.0-0.nightly-2024-09-29-173421 True False 13m Cluster version is 4.17.0-0.nightly-2024-09-29-173421' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/sidecar/render.sh ++ export SUITE_DIR=./tests/e2e/sidecar ++ SUITE_DIR=./tests/e2e/sidecar ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/sidecar ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + jaeger_service_name=order + start_test sidecar-deployment + '[' 1 -ne 1 ']' + test_name=sidecar-deployment + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-deployment' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-deployment\e[0m' Rendering files for test sidecar-deployment + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build + '[' _build '!=' _build ']' + mkdir -p sidecar-deployment + cd sidecar-deployment + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-namespace + '[' 1 -ne 1 ']' + test_name=sidecar-namespace + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-namespace' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-namespace\e[0m' Rendering files for test sidecar-namespace + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-deployment + '[' sidecar-deployment '!=' _build ']' + cd .. + mkdir -p sidecar-namespace + cd sidecar-namespace + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-skip-webhook + '[' 1 -ne 1 ']' + test_name=sidecar-skip-webhook + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-skip-webhook' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-skip-webhook\e[0m' Rendering files for test sidecar-skip-webhook + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-namespace + '[' sidecar-namespace '!=' _build ']' + cd .. + mkdir -p sidecar-skip-webhook + cd sidecar-skip-webhook + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running sidecar E2E tests' Running sidecar E2E tests + cd tests/e2e/sidecar/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2856073267 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 4 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/sidecar-deployment === PAUSE kuttl/harness/sidecar-deployment === RUN kuttl/harness/sidecar-namespace === PAUSE kuttl/harness/sidecar-namespace === RUN kuttl/harness/sidecar-skip-webhook === PAUSE kuttl/harness/sidecar-skip-webhook === CONT kuttl/harness/artifacts logger.go:42: 17:51:02 | artifacts | Creating namespace: kuttl-test-exact-walrus logger.go:42: 17:51:02 | artifacts | artifacts events from ns kuttl-test-exact-walrus: logger.go:42: 17:51:02 | artifacts | Deleting namespace: kuttl-test-exact-walrus === CONT kuttl/harness/sidecar-deployment logger.go:42: 17:51:08 | sidecar-deployment | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 17:51:08 | sidecar-deployment | Creating namespace: kuttl-test-perfect-bass logger.go:42: 17:51:08 | sidecar-deployment/0-install | starting test step 0-install logger.go:42: 17:51:08 | sidecar-deployment/0-install | Jaeger:kuttl-test-perfect-bass/agent-as-sidecar created logger.go:42: 17:51:18 | sidecar-deployment/0-install | test step completed 0-install logger.go:42: 17:51:18 | sidecar-deployment/1-install | starting test step 1-install logger.go:42: 17:51:18 | sidecar-deployment/1-install | Deployment:kuttl-test-perfect-bass/vertx-create-span-sidecar created logger.go:42: 17:51:23 | sidecar-deployment/1-install | test step completed 1-install logger.go:42: 17:51:23 | sidecar-deployment/2-enable-injection | starting test step 2-enable-injection logger.go:42: 17:51:23 | sidecar-deployment/2-enable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-perfect-bass] logger.go:42: 17:51:24 | sidecar-deployment/2-enable-injection | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 17:51:32 | sidecar-deployment/2-enable-injection | test step completed 2-enable-injection logger.go:42: 17:51:32 | sidecar-deployment/3-find-service | starting test step 3-find-service logger.go:42: 17:51:32 | sidecar-deployment/3-find-service | Job:kuttl-test-perfect-bass/00-find-service created logger.go:42: 17:51:45 | sidecar-deployment/3-find-service | test step completed 3-find-service logger.go:42: 17:51:45 | sidecar-deployment/4-other-instance | starting test step 4-other-instance logger.go:42: 17:51:45 | sidecar-deployment/4-other-instance | Jaeger:kuttl-test-perfect-bass/agent-as-sidecar2 created logger.go:42: 17:51:52 | sidecar-deployment/4-other-instance | test step completed 4-other-instance logger.go:42: 17:51:52 | sidecar-deployment/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 17:51:52 | sidecar-deployment/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 17:51:52 | sidecar-deployment/6-find-service | starting test step 6-find-service logger.go:42: 17:51:52 | sidecar-deployment/6-find-service | Job:kuttl-test-perfect-bass/01-find-service created logger.go:42: 17:52:13 | sidecar-deployment/6-find-service | test step completed 6-find-service logger.go:42: 17:52:13 | sidecar-deployment/7-disable-injection | starting test step 7-disable-injection logger.go:42: 17:52:13 | sidecar-deployment/7-disable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=false --namespace kuttl-test-perfect-bass] logger.go:42: 17:52:13 | sidecar-deployment/7-disable-injection | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 17:52:16 | sidecar-deployment/7-disable-injection | test step completed 7-disable-injection logger.go:42: 17:52:16 | sidecar-deployment | sidecar-deployment events from ns kuttl-test-perfect-bass: logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:11 +0000 UTC Normal Pod agent-as-sidecar-77857d66c6-zsvjd Binding Scheduled Successfully assigned kuttl-test-perfect-bass/agent-as-sidecar-77857d66c6-zsvjd to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:11 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-77857d66c6 SuccessfulCreate Created pod: agent-as-sidecar-77857d66c6-zsvjd replicaset-controller logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:11 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-77857d66c6 to 1 deployment-controller logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:12 +0000 UTC Normal Pod agent-as-sidecar-77857d66c6-zsvjd AddedInterface Add eth0 [10.131.0.29/23] from ovn-kubernetes multus logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:12 +0000 UTC Normal Pod agent-as-sidecar-77857d66c6-zsvjd.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:16 +0000 UTC Normal Pod agent-as-sidecar-77857d66c6-zsvjd.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" in 3.652s (3.652s including waiting). Image size: 147237792 bytes. kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:16 +0000 UTC Normal Pod agent-as-sidecar-77857d66c6-zsvjd.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:16 +0000 UTC Normal Pod agent-as-sidecar-77857d66c6-zsvjd.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:18 +0000 UTC Normal Pod vertx-create-span-sidecar-79989769b8-9mq72 Binding Scheduled Successfully assigned kuttl-test-perfect-bass/vertx-create-span-sidecar-79989769b8-9mq72 to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:18 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-79989769b8 SuccessfulCreate Created pod: vertx-create-span-sidecar-79989769b8-9mq72 replicaset-controller logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:18 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-79989769b8 to 1 deployment-controller logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:19 +0000 UTC Normal Pod vertx-create-span-sidecar-79989769b8-9mq72 AddedInterface Add eth0 [10.129.2.22/23] from ovn-kubernetes multus logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:19 +0000 UTC Normal Pod vertx-create-span-sidecar-79989769b8-9mq72.spec.containers{vertx-create-span-sidecar} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:23 +0000 UTC Normal Pod vertx-create-span-sidecar-79989769b8-9mq72.spec.containers{vertx-create-span-sidecar} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 3.704s (3.704s including waiting). Image size: 282912835 bytes. kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:23 +0000 UTC Normal Pod vertx-create-span-sidecar-79989769b8-9mq72.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:23 +0000 UTC Normal Pod vertx-create-span-sidecar-79989769b8-9mq72.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:24 +0000 UTC Normal Pod vertx-create-span-sidecar-6bdf4c9f96-9x7wf Binding Scheduled Successfully assigned kuttl-test-perfect-bass/vertx-create-span-sidecar-6bdf4c9f96-9x7wf to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:24 +0000 UTC Normal Pod vertx-create-span-sidecar-6bdf4c9f96-9x7wf AddedInterface Add eth0 [10.128.2.16/23] from ovn-kubernetes multus logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:24 +0000 UTC Normal Pod vertx-create-span-sidecar-6bdf4c9f96-9x7wf.spec.containers{vertx-create-span-sidecar} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:24 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-6bdf4c9f96 SuccessfulCreate Created pod: vertx-create-span-sidecar-6bdf4c9f96-9x7wf replicaset-controller logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:24 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-6bdf4c9f96 to 1 deployment-controller logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:29 +0000 UTC Normal Pod vertx-create-span-sidecar-6bdf4c9f96-9x7wf.spec.containers{vertx-create-span-sidecar} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 4.33s (4.33s including waiting). Image size: 282912835 bytes. kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:29 +0000 UTC Normal Pod vertx-create-span-sidecar-6bdf4c9f96-9x7wf.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:29 +0000 UTC Normal Pod vertx-create-span-sidecar-6bdf4c9f96-9x7wf.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:29 +0000 UTC Normal Pod vertx-create-span-sidecar-6bdf4c9f96-9x7wf.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:31 +0000 UTC Normal Pod vertx-create-span-sidecar-6bdf4c9f96-9x7wf.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" in 2.028s (2.028s including waiting). Image size: 115523131 bytes. kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:31 +0000 UTC Normal Pod vertx-create-span-sidecar-6bdf4c9f96-9x7wf.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:31 +0000 UTC Normal Pod vertx-create-span-sidecar-6bdf4c9f96-9x7wf.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:32 +0000 UTC Normal Pod 00-find-service-xbgnt Binding Scheduled Successfully assigned kuttl-test-perfect-bass/00-find-service-xbgnt to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:32 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-xbgnt job-controller logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:32 +0000 UTC Warning Pod vertx-create-span-sidecar-79989769b8-9mq72.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.22:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:32 +0000 UTC Warning Pod vertx-create-span-sidecar-79989769b8-9mq72.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.22:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:33 +0000 UTC Normal Pod 00-find-service-xbgnt AddedInterface Add eth0 [10.131.0.30/23] from ovn-kubernetes multus logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:33 +0000 UTC Normal Pod 00-find-service-xbgnt.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:34 +0000 UTC Normal Pod 00-find-service-xbgnt.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.581s (1.581s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:34 +0000 UTC Normal Pod 00-find-service-xbgnt.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:34 +0000 UTC Normal Pod 00-find-service-xbgnt.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:34 +0000 UTC Normal Pod vertx-create-span-sidecar-79989769b8-9mq72.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:34 +0000 UTC Warning Pod vertx-create-span-sidecar-79989769b8-9mq72.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.22:8080/": read tcp 10.129.2.2:33000->10.129.2.22:8080: read: connection reset by peer kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:34 +0000 UTC Warning Pod vertx-create-span-sidecar-79989769b8-9mq72.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.22:8080/": dial tcp 10.129.2.22:8080: connect: connection refused kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:34 +0000 UTC Normal Pod vertx-create-span-sidecar-79989769b8-9mq72.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:37 +0000 UTC Warning Pod vertx-create-span-sidecar-6bdf4c9f96-9x7wf.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.128.2.16:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:37 +0000 UTC Warning Pod vertx-create-span-sidecar-6bdf4c9f96-9x7wf.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.16:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:39 +0000 UTC Normal Pod vertx-create-span-sidecar-6bdf4c9f96-9x7wf.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:39 +0000 UTC Warning Pod vertx-create-span-sidecar-6bdf4c9f96-9x7wf.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.16:8080/": read tcp 10.128.2.2:47352->10.128.2.16:8080: read: connection reset by peer kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:39 +0000 UTC Warning Pod vertx-create-span-sidecar-6bdf4c9f96-9x7wf.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.16:8080/": dial tcp 10.128.2.16:8080: connect: connection refused kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:40 +0000 UTC Normal Pod vertx-create-span-sidecar-6bdf4c9f96-9x7wf.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:45 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:45 +0000 UTC Warning Pod vertx-create-span-sidecar-79989769b8-9mq72.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.22:8080/": read tcp 10.129.2.2:50348->10.129.2.22:8080: read: connection reset by peer kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:49 +0000 UTC Normal Pod agent-as-sidecar2-8b45cb664-5zc5q Binding Scheduled Successfully assigned kuttl-test-perfect-bass/agent-as-sidecar2-8b45cb664-5zc5q to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:49 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-8b45cb664 SuccessfulCreate Created pod: agent-as-sidecar2-8b45cb664-5zc5q replicaset-controller logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:49 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-8b45cb664 to 1 deployment-controller logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:50 +0000 UTC Normal Pod agent-as-sidecar2-8b45cb664-5zc5q AddedInterface Add eth0 [10.131.0.31/23] from ovn-kubernetes multus logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:50 +0000 UTC Normal Pod agent-as-sidecar2-8b45cb664-5zc5q.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:50 +0000 UTC Normal Pod agent-as-sidecar2-8b45cb664-5zc5q.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:50 +0000 UTC Normal Pod agent-as-sidecar2-8b45cb664-5zc5q.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:52 +0000 UTC Normal Pod 01-find-service-vtdk9 Binding Scheduled Successfully assigned kuttl-test-perfect-bass/01-find-service-vtdk9 to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:52 +0000 UTC Normal Pod 01-find-service-vtdk9 AddedInterface Add eth0 [10.129.2.23/23] from ovn-kubernetes multus logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:52 +0000 UTC Normal Pod 01-find-service-vtdk9.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:52 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-vtdk9 job-controller logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:52 +0000 UTC Normal Pod agent-as-sidecar-77857d66c6-zsvjd.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:52 +0000 UTC Warning Pod agent-as-sidecar-77857d66c6-zsvjd.spec.containers{jaeger} Unhealthy Readiness probe failed: Get "http://10.131.0.29:14269/": dial tcp 10.131.0.29:14269: connect: connection refused kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:54 +0000 UTC Normal Pod 01-find-service-vtdk9.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.326s (1.326s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:54 +0000 UTC Normal Pod 01-find-service-vtdk9.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:54 +0000 UTC Normal Pod 01-find-service-vtdk9.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:55 +0000 UTC Normal Pod vertx-create-span-sidecar-5b7c8b44c6-c78fr Binding Scheduled Successfully assigned kuttl-test-perfect-bass/vertx-create-span-sidecar-5b7c8b44c6-c78fr to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:55 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-5b7c8b44c6 SuccessfulCreate Created pod: vertx-create-span-sidecar-5b7c8b44c6-c78fr replicaset-controller logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:55 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-79989769b8 SuccessfulDelete Deleted pod: vertx-create-span-sidecar-79989769b8-9mq72 replicaset-controller logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:55 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-79989769b8 to 0 from 1 deployment-controller logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:55 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-5b7c8b44c6 to 1 from 0 deployment-controller logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:56 +0000 UTC Normal Pod vertx-create-span-sidecar-5b7c8b44c6-c78fr AddedInterface Add eth0 [10.131.0.32/23] from ovn-kubernetes multus logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:56 +0000 UTC Normal Pod vertx-create-span-sidecar-5b7c8b44c6-c78fr.spec.containers{vertx-create-span-sidecar} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:59 +0000 UTC Normal Pod vertx-create-span-sidecar-5b7c8b44c6-c78fr.spec.containers{vertx-create-span-sidecar} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 3.234s (3.234s including waiting). Image size: 282912835 bytes. kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:59 +0000 UTC Normal Pod vertx-create-span-sidecar-5b7c8b44c6-c78fr.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:59 +0000 UTC Normal Pod vertx-create-span-sidecar-5b7c8b44c6-c78fr.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:59 +0000 UTC Normal Pod vertx-create-span-sidecar-5b7c8b44c6-c78fr.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:59 +0000 UTC Normal Pod vertx-create-span-sidecar-5b7c8b44c6-c78fr.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:51:59 +0000 UTC Normal Pod vertx-create-span-sidecar-5b7c8b44c6-c78fr.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:52:07 +0000 UTC Warning Pod vertx-create-span-sidecar-5b7c8b44c6-c78fr.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.131.0.32:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:52:07 +0000 UTC Warning Pod vertx-create-span-sidecar-5b7c8b44c6-c78fr.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.32:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:52:10 +0000 UTC Normal Pod vertx-create-span-sidecar-5b7c8b44c6-c78fr.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:52:10 +0000 UTC Warning Pod vertx-create-span-sidecar-5b7c8b44c6-c78fr.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.32:8080/": read tcp 10.131.0.2:40520->10.131.0.32:8080: read: connection reset by peer kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:52:10 +0000 UTC Warning Pod vertx-create-span-sidecar-5b7c8b44c6-c78fr.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.32:8080/": dial tcp 10.131.0.32:8080: connect: connection refused kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:52:10 +0000 UTC Normal Pod vertx-create-span-sidecar-5b7c8b44c6-c78fr.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:52:12 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:52:13 +0000 UTC Normal Pod vertx-create-span-sidecar-66dc5d774d-vvqr7 Binding Scheduled Successfully assigned kuttl-test-perfect-bass/vertx-create-span-sidecar-66dc5d774d-vvqr7 to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:52:13 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-66dc5d774d SuccessfulCreate Created pod: vertx-create-span-sidecar-66dc5d774d-vvqr7 replicaset-controller logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:52:13 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-6bdf4c9f96 SuccessfulDelete Deleted pod: vertx-create-span-sidecar-6bdf4c9f96-9x7wf replicaset-controller logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:52:13 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-6bdf4c9f96 to 0 from 1 deployment-controller logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:52:13 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-66dc5d774d to 1 from 0 deployment-controller logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:52:14 +0000 UTC Normal Pod vertx-create-span-sidecar-66dc5d774d-vvqr7 AddedInterface Add eth0 [10.129.2.24/23] from ovn-kubernetes multus logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:52:14 +0000 UTC Normal Pod vertx-create-span-sidecar-66dc5d774d-vvqr7.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:52:14 +0000 UTC Normal Pod vertx-create-span-sidecar-66dc5d774d-vvqr7.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 17:52:16 | sidecar-deployment | 2024-09-30 17:52:14 +0000 UTC Normal Pod vertx-create-span-sidecar-66dc5d774d-vvqr7.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 17:52:16 | sidecar-deployment | Deleting namespace: kuttl-test-perfect-bass === CONT kuttl/harness/sidecar-namespace logger.go:42: 17:52:22 | sidecar-namespace | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 17:52:22 | sidecar-namespace | Creating namespace: kuttl-test-living-grubworm logger.go:42: 17:52:22 | sidecar-namespace/0-install | starting test step 0-install logger.go:42: 17:52:22 | sidecar-namespace/0-install | Jaeger:kuttl-test-living-grubworm/agent-as-sidecar created logger.go:42: 17:52:33 | sidecar-namespace/0-install | test step completed 0-install logger.go:42: 17:52:33 | sidecar-namespace/1-install | starting test step 1-install logger.go:42: 17:52:33 | sidecar-namespace/1-install | Deployment:kuttl-test-living-grubworm/vertx-create-span-sidecar created logger.go:42: 17:52:35 | sidecar-namespace/1-install | test step completed 1-install logger.go:42: 17:52:35 | sidecar-namespace/2-enable-injection | starting test step 2-enable-injection logger.go:42: 17:52:35 | sidecar-namespace/2-enable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="true"] logger.go:42: 17:52:35 | sidecar-namespace/2-enable-injection | namespace/kuttl-test-living-grubworm annotated logger.go:42: 17:52:37 | sidecar-namespace/2-enable-injection | test step completed 2-enable-injection logger.go:42: 17:52:37 | sidecar-namespace/3-find-service | starting test step 3-find-service logger.go:42: 17:52:37 | sidecar-namespace/3-find-service | Job:kuttl-test-living-grubworm/00-find-service created logger.go:42: 17:52:51 | sidecar-namespace/3-find-service | test step completed 3-find-service logger.go:42: 17:52:51 | sidecar-namespace/4-other-instance | starting test step 4-other-instance logger.go:42: 17:52:51 | sidecar-namespace/4-other-instance | Jaeger:kuttl-test-living-grubworm/agent-as-sidecar2 created logger.go:42: 17:53:02 | sidecar-namespace/4-other-instance | test step completed 4-other-instance logger.go:42: 17:53:02 | sidecar-namespace/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 17:53:02 | sidecar-namespace/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 17:53:02 | sidecar-namespace/6-find-service | starting test step 6-find-service logger.go:42: 17:53:02 | sidecar-namespace/6-find-service | Job:kuttl-test-living-grubworm/01-find-service created logger.go:42: 17:53:21 | sidecar-namespace/6-find-service | test step completed 6-find-service logger.go:42: 17:53:21 | sidecar-namespace/7-disable-injection | starting test step 7-disable-injection logger.go:42: 17:53:21 | sidecar-namespace/7-disable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="false"] logger.go:42: 17:53:22 | sidecar-namespace/7-disable-injection | namespace/kuttl-test-living-grubworm annotated logger.go:42: 17:53:24 | sidecar-namespace/7-disable-injection | test step completed 7-disable-injection logger.go:42: 17:53:24 | sidecar-namespace | sidecar-namespace events from ns kuttl-test-living-grubworm: logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:26 +0000 UTC Normal Pod agent-as-sidecar-684ddc9556-6p5sq Binding Scheduled Successfully assigned kuttl-test-living-grubworm/agent-as-sidecar-684ddc9556-6p5sq to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:26 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-684ddc9556 SuccessfulCreate Created pod: agent-as-sidecar-684ddc9556-6p5sq replicaset-controller logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:26 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-684ddc9556 to 1 deployment-controller logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:27 +0000 UTC Normal Pod agent-as-sidecar-684ddc9556-6p5sq AddedInterface Add eth0 [10.129.2.25/23] from ovn-kubernetes multus logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:27 +0000 UTC Normal Pod agent-as-sidecar-684ddc9556-6p5sq.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:30 +0000 UTC Normal Pod agent-as-sidecar-684ddc9556-6p5sq.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" in 3.629s (3.629s including waiting). Image size: 147237792 bytes. kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:31 +0000 UTC Normal Pod agent-as-sidecar-684ddc9556-6p5sq.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:31 +0000 UTC Normal Pod agent-as-sidecar-684ddc9556-6p5sq.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:33 +0000 UTC Normal Pod vertx-create-span-sidecar-79989769b8-nxtnl Binding Scheduled Successfully assigned kuttl-test-living-grubworm/vertx-create-span-sidecar-79989769b8-nxtnl to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:33 +0000 UTC Normal Pod vertx-create-span-sidecar-79989769b8-nxtnl AddedInterface Add eth0 [10.128.2.17/23] from ovn-kubernetes multus logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:33 +0000 UTC Normal Pod vertx-create-span-sidecar-79989769b8-nxtnl.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:33 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-79989769b8 SuccessfulCreate Created pod: vertx-create-span-sidecar-79989769b8-nxtnl replicaset-controller logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:33 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-79989769b8 to 1 deployment-controller logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:34 +0000 UTC Normal Pod vertx-create-span-sidecar-79989769b8-nxtnl.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:34 +0000 UTC Normal Pod vertx-create-span-sidecar-79989769b8-nxtnl.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:35 +0000 UTC Normal Pod vertx-create-span-sidecar-6468b744b-mj9lq Binding Scheduled Successfully assigned kuttl-test-living-grubworm/vertx-create-span-sidecar-6468b744b-mj9lq to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:35 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-6468b744b SuccessfulCreate Created pod: vertx-create-span-sidecar-6468b744b-mj9lq replicaset-controller logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:35 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-6468b744b to 1 deployment-controller logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:36 +0000 UTC Normal Pod vertx-create-span-sidecar-6468b744b-mj9lq AddedInterface Add eth0 [10.131.0.33/23] from ovn-kubernetes multus logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:36 +0000 UTC Normal Pod vertx-create-span-sidecar-6468b744b-mj9lq.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:36 +0000 UTC Normal Pod vertx-create-span-sidecar-6468b744b-mj9lq.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:36 +0000 UTC Normal Pod vertx-create-span-sidecar-6468b744b-mj9lq.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:36 +0000 UTC Normal Pod vertx-create-span-sidecar-6468b744b-mj9lq.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:36 +0000 UTC Normal Pod vertx-create-span-sidecar-6468b744b-mj9lq.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:36 +0000 UTC Normal Pod vertx-create-span-sidecar-6468b744b-mj9lq.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:37 +0000 UTC Normal Pod 00-find-service-9hstk Binding Scheduled Successfully assigned kuttl-test-living-grubworm/00-find-service-9hstk to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:37 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-9hstk job-controller logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:38 +0000 UTC Normal Pod 00-find-service-9hstk AddedInterface Add eth0 [10.128.2.18/23] from ovn-kubernetes multus logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:38 +0000 UTC Normal Pod 00-find-service-9hstk.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:39 +0000 UTC Normal Pod 00-find-service-9hstk.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.469s (1.469s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:39 +0000 UTC Normal Pod 00-find-service-9hstk.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:40 +0000 UTC Normal Pod 00-find-service-9hstk.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:42 +0000 UTC Warning Pod vertx-create-span-sidecar-79989769b8-nxtnl.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.17:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:42 +0000 UTC Warning Pod vertx-create-span-sidecar-79989769b8-nxtnl.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.128.2.17:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:44 +0000 UTC Warning Pod vertx-create-span-sidecar-6468b744b-mj9lq.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.131.0.33:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:44 +0000 UTC Warning Pod vertx-create-span-sidecar-6468b744b-mj9lq.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.33:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:44 +0000 UTC Normal Pod vertx-create-span-sidecar-79989769b8-nxtnl.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:45 +0000 UTC Warning Pod vertx-create-span-sidecar-79989769b8-nxtnl.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.17:8080/": read tcp 10.128.2.2:52146->10.128.2.17:8080: read: connection reset by peer kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:45 +0000 UTC Warning Pod vertx-create-span-sidecar-79989769b8-nxtnl.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.17:8080/": dial tcp 10.128.2.17:8080: connect: connection refused kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:47 +0000 UTC Normal Pod vertx-create-span-sidecar-6468b744b-mj9lq.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:47 +0000 UTC Warning Pod vertx-create-span-sidecar-6468b744b-mj9lq.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.33:8080/": read tcp 10.131.0.2:41306->10.131.0.33:8080: read: connection reset by peer kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:47 +0000 UTC Warning Pod vertx-create-span-sidecar-6468b744b-mj9lq.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.33:8080/": dial tcp 10.131.0.33:8080: connect: connection refused kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:50 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:56 +0000 UTC Warning Pod vertx-create-span-sidecar-79989769b8-nxtnl.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.17:8080/": read tcp 10.128.2.2:47128->10.128.2.17:8080: read: connection reset by peer kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:57 +0000 UTC Normal Pod agent-as-sidecar2-6b75d4f7d4-m4v79 Binding Scheduled Successfully assigned kuttl-test-living-grubworm/agent-as-sidecar2-6b75d4f7d4-m4v79 to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:57 +0000 UTC Normal Pod agent-as-sidecar2-6b75d4f7d4-m4v79 AddedInterface Add eth0 [10.128.2.19/23] from ovn-kubernetes multus logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:57 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-6b75d4f7d4 SuccessfulCreate Created pod: agent-as-sidecar2-6b75d4f7d4-m4v79 replicaset-controller logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:57 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-6b75d4f7d4 to 1 deployment-controller logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:52:58 +0000 UTC Normal Pod agent-as-sidecar2-6b75d4f7d4-m4v79.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:00 +0000 UTC Normal Pod agent-as-sidecar2-6b75d4f7d4-m4v79.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" in 2.542s (2.542s including waiting). Image size: 147237792 bytes. kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:00 +0000 UTC Normal Pod agent-as-sidecar2-6b75d4f7d4-m4v79.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:00 +0000 UTC Normal Pod agent-as-sidecar2-6b75d4f7d4-m4v79.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:02 +0000 UTC Normal Pod 01-find-service-bjtlm Binding Scheduled Successfully assigned kuttl-test-living-grubworm/01-find-service-bjtlm to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:02 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-bjtlm job-controller logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:02 +0000 UTC Normal Pod agent-as-sidecar-684ddc9556-6p5sq.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:03 +0000 UTC Normal Pod 01-find-service-bjtlm AddedInterface Add eth0 [10.129.2.26/23] from ovn-kubernetes multus logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:03 +0000 UTC Normal Pod 01-find-service-bjtlm.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:03 +0000 UTC Normal Pod 01-find-service-bjtlm.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 208ms (208ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:03 +0000 UTC Normal Pod 01-find-service-bjtlm.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:03 +0000 UTC Normal Pod 01-find-service-bjtlm.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:05 +0000 UTC Normal Pod vertx-create-span-sidecar-664c987c8-6hdwk Binding Scheduled Successfully assigned kuttl-test-living-grubworm/vertx-create-span-sidecar-664c987c8-6hdwk to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:05 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-664c987c8 SuccessfulCreate Created pod: vertx-create-span-sidecar-664c987c8-6hdwk replicaset-controller logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:05 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-79989769b8 SuccessfulDelete Deleted pod: vertx-create-span-sidecar-79989769b8-nxtnl replicaset-controller logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:05 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-79989769b8 to 0 from 1 deployment-controller logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:05 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-664c987c8 to 1 from 0 deployment-controller logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:06 +0000 UTC Normal Pod vertx-create-span-sidecar-664c987c8-6hdwk AddedInterface Add eth0 [10.129.2.27/23] from ovn-kubernetes multus logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:06 +0000 UTC Normal Pod vertx-create-span-sidecar-664c987c8-6hdwk.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:06 +0000 UTC Normal Pod vertx-create-span-sidecar-664c987c8-6hdwk.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:06 +0000 UTC Normal Pod vertx-create-span-sidecar-664c987c8-6hdwk.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:06 +0000 UTC Normal Pod vertx-create-span-sidecar-664c987c8-6hdwk.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:09 +0000 UTC Normal Pod vertx-create-span-sidecar-664c987c8-6hdwk.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" in 2.908s (2.908s including waiting). Image size: 115523131 bytes. kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:09 +0000 UTC Normal Pod vertx-create-span-sidecar-664c987c8-6hdwk.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:09 +0000 UTC Normal Pod vertx-create-span-sidecar-664c987c8-6hdwk.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:14 +0000 UTC Warning Pod vertx-create-span-sidecar-664c987c8-6hdwk.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.27:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:14 +0000 UTC Warning Pod vertx-create-span-sidecar-664c987c8-6hdwk.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.27:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:17 +0000 UTC Normal Pod vertx-create-span-sidecar-664c987c8-6hdwk.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:17 +0000 UTC Warning Pod vertx-create-span-sidecar-664c987c8-6hdwk.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.27:8080/": read tcp 10.129.2.2:45842->10.129.2.27:8080: read: connection reset by peer kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:17 +0000 UTC Warning Pod vertx-create-span-sidecar-664c987c8-6hdwk.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.27:8080/": dial tcp 10.129.2.27:8080: connect: connection refused kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:21 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:22 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-6468b744b SuccessfulDelete Deleted pod: vertx-create-span-sidecar-6468b744b-mj9lq replicaset-controller logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:22 +0000 UTC Normal Pod vertx-create-span-sidecar-67c4cd6584-d7nh5 Binding Scheduled Successfully assigned kuttl-test-living-grubworm/vertx-create-span-sidecar-67c4cd6584-d7nh5 to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:22 +0000 UTC Normal Pod vertx-create-span-sidecar-67c4cd6584-d7nh5 AddedInterface Add eth0 [10.128.2.20/23] from ovn-kubernetes multus logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:22 +0000 UTC Normal Pod vertx-create-span-sidecar-67c4cd6584-d7nh5.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:22 +0000 UTC Normal Pod vertx-create-span-sidecar-67c4cd6584-d7nh5.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:22 +0000 UTC Normal Pod vertx-create-span-sidecar-67c4cd6584-d7nh5.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:22 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-67c4cd6584 SuccessfulCreate Created pod: vertx-create-span-sidecar-67c4cd6584-d7nh5 replicaset-controller logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:22 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-6468b744b to 0 from 1 deployment-controller logger.go:42: 17:53:24 | sidecar-namespace | 2024-09-30 17:53:22 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-67c4cd6584 to 1 from 0 deployment-controller logger.go:42: 17:53:24 | sidecar-namespace | Deleting namespace: kuttl-test-living-grubworm === CONT kuttl/harness/sidecar-skip-webhook logger.go:42: 17:53:31 | sidecar-skip-webhook | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 17:53:31 | sidecar-skip-webhook | Creating namespace: kuttl-test-grateful-kitten logger.go:42: 17:53:31 | sidecar-skip-webhook/0-install | starting test step 0-install logger.go:42: 17:53:31 | sidecar-skip-webhook/0-install | Jaeger:kuttl-test-grateful-kitten/agent-as-sidecar created logger.go:42: 17:53:37 | sidecar-skip-webhook/0-install | test step completed 0-install logger.go:42: 17:53:37 | sidecar-skip-webhook/1-install | starting test step 1-install logger.go:42: 17:53:37 | sidecar-skip-webhook/1-install | Deployment:kuttl-test-grateful-kitten/vertx-create-span-sidecar created logger.go:42: 17:53:39 | sidecar-skip-webhook/1-install | test step completed 1-install logger.go:42: 17:53:39 | sidecar-skip-webhook/2-add-anotation-and-label | starting test step 2-add-anotation-and-label logger.go:42: 17:53:39 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name=jaeger-operator --namespace kuttl-test-grateful-kitten] logger.go:42: 17:53:39 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar labeled logger.go:42: 17:53:39 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-grateful-kitten] logger.go:42: 17:53:39 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 17:53:40 | sidecar-skip-webhook/2-add-anotation-and-label | test step completed 2-add-anotation-and-label logger.go:42: 17:53:40 | sidecar-skip-webhook/3-remove-label | starting test step 3-remove-label logger.go:42: 17:53:40 | sidecar-skip-webhook/3-remove-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name- --namespace kuttl-test-grateful-kitten] logger.go:42: 17:53:40 | sidecar-skip-webhook/3-remove-label | deployment.apps/vertx-create-span-sidecar unlabeled logger.go:42: 17:53:42 | sidecar-skip-webhook/3-remove-label | test step completed 3-remove-label logger.go:42: 17:53:42 | sidecar-skip-webhook | sidecar-skip-webhook events from ns kuttl-test-grateful-kitten: logger.go:42: 17:53:42 | sidecar-skip-webhook | 2024-09-30 17:53:34 +0000 UTC Normal Pod agent-as-sidecar-677b67dd7c-fx4ng Binding Scheduled Successfully assigned kuttl-test-grateful-kitten/agent-as-sidecar-677b67dd7c-fx4ng to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 17:53:42 | sidecar-skip-webhook | 2024-09-30 17:53:34 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-677b67dd7c SuccessfulCreate Created pod: agent-as-sidecar-677b67dd7c-fx4ng replicaset-controller logger.go:42: 17:53:42 | sidecar-skip-webhook | 2024-09-30 17:53:34 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-677b67dd7c to 1 deployment-controller logger.go:42: 17:53:42 | sidecar-skip-webhook | 2024-09-30 17:53:35 +0000 UTC Normal Pod agent-as-sidecar-677b67dd7c-fx4ng AddedInterface Add eth0 [10.128.2.21/23] from ovn-kubernetes multus logger.go:42: 17:53:42 | sidecar-skip-webhook | 2024-09-30 17:53:35 +0000 UTC Normal Pod agent-as-sidecar-677b67dd7c-fx4ng.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 17:53:42 | sidecar-skip-webhook | 2024-09-30 17:53:35 +0000 UTC Normal Pod agent-as-sidecar-677b67dd7c-fx4ng.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 17:53:42 | sidecar-skip-webhook | 2024-09-30 17:53:35 +0000 UTC Normal Pod agent-as-sidecar-677b67dd7c-fx4ng.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 17:53:42 | sidecar-skip-webhook | 2024-09-30 17:53:37 +0000 UTC Normal Pod vertx-create-span-sidecar-79989769b8-cv5b7 Binding Scheduled Successfully assigned kuttl-test-grateful-kitten/vertx-create-span-sidecar-79989769b8-cv5b7 to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 17:53:42 | sidecar-skip-webhook | 2024-09-30 17:53:37 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-79989769b8 SuccessfulCreate Created pod: vertx-create-span-sidecar-79989769b8-cv5b7 replicaset-controller logger.go:42: 17:53:42 | sidecar-skip-webhook | 2024-09-30 17:53:37 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-79989769b8 to 1 deployment-controller logger.go:42: 17:53:42 | sidecar-skip-webhook | 2024-09-30 17:53:38 +0000 UTC Normal Pod vertx-create-span-sidecar-79989769b8-cv5b7 AddedInterface Add eth0 [10.129.2.28/23] from ovn-kubernetes multus logger.go:42: 17:53:42 | sidecar-skip-webhook | 2024-09-30 17:53:38 +0000 UTC Normal Pod vertx-create-span-sidecar-79989769b8-cv5b7.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 17:53:42 | sidecar-skip-webhook | 2024-09-30 17:53:38 +0000 UTC Normal Pod vertx-create-span-sidecar-79989769b8-cv5b7.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 17:53:42 | sidecar-skip-webhook | 2024-09-30 17:53:38 +0000 UTC Normal Pod vertx-create-span-sidecar-79989769b8-cv5b7.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 17:53:42 | sidecar-skip-webhook | 2024-09-30 17:53:39 +0000 UTC Normal Pod vertx-create-span-sidecar-77b656b799-ctxvn Binding Scheduled Successfully assigned kuttl-test-grateful-kitten/vertx-create-span-sidecar-77b656b799-ctxvn to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 17:53:42 | sidecar-skip-webhook | 2024-09-30 17:53:39 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-77b656b799 SuccessfulCreate Created pod: vertx-create-span-sidecar-77b656b799-ctxvn replicaset-controller logger.go:42: 17:53:42 | sidecar-skip-webhook | 2024-09-30 17:53:39 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-77b656b799 to 1 deployment-controller logger.go:42: 17:53:42 | sidecar-skip-webhook | 2024-09-30 17:53:40 +0000 UTC Normal Pod vertx-create-span-sidecar-77b656b799-ctxvn AddedInterface Add eth0 [10.131.0.34/23] from ovn-kubernetes multus logger.go:42: 17:53:42 | sidecar-skip-webhook | 2024-09-30 17:53:40 +0000 UTC Normal Pod vertx-create-span-sidecar-77b656b799-ctxvn.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 17:53:42 | sidecar-skip-webhook | 2024-09-30 17:53:40 +0000 UTC Normal Pod vertx-create-span-sidecar-77b656b799-ctxvn.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 17:53:42 | sidecar-skip-webhook | 2024-09-30 17:53:40 +0000 UTC Normal Pod vertx-create-span-sidecar-77b656b799-ctxvn.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 17:53:42 | sidecar-skip-webhook | 2024-09-30 17:53:40 +0000 UTC Normal Pod vertx-create-span-sidecar-77b656b799-ctxvn.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 17:53:42 | sidecar-skip-webhook | 2024-09-30 17:53:40 +0000 UTC Normal Pod vertx-create-span-sidecar-77b656b799-ctxvn.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 17:53:42 | sidecar-skip-webhook | 2024-09-30 17:53:40 +0000 UTC Normal Pod vertx-create-span-sidecar-77b656b799-ctxvn.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 17:53:42 | sidecar-skip-webhook | Deleting namespace: kuttl-test-grateful-kitten === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (166.43s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.90s) --- PASS: kuttl/harness/sidecar-deployment (74.61s) --- PASS: kuttl/harness/sidecar-namespace (68.36s) --- PASS: kuttl/harness/sidecar-skip-webhook (17.37s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name sidecar --report --output /logs/artifacts/sidecar.xml ./artifacts/kuttl-report.xml time="2024-09-30T17:53:48Z" level=debug msg="Setting a new name for the test suites" time="2024-09-30T17:53:48Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-09-30T17:53:48Z" level=debug msg="normalizing test case names" time="2024-09-30T17:53:48Z" level=debug msg="sidecar/artifacts -> sidecar_artifacts" time="2024-09-30T17:53:48Z" level=debug msg="sidecar/sidecar-deployment -> sidecar_sidecar_deployment" time="2024-09-30T17:53:48Z" level=debug msg="sidecar/sidecar-namespace -> sidecar_sidecar_namespace" time="2024-09-30T17:53:48Z" level=debug msg="sidecar/sidecar-skip-webhook -> sidecar_sidecar_skip_webhook" +------------------------------+--------+ | NAME | RESULT | +------------------------------+--------+ | sidecar_artifacts | passed | | sidecar_sidecar_deployment | passed | | sidecar_sidecar_namespace | passed | | sidecar_sidecar_skip_webhook | passed | +------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh ui false true + '[' 3 -ne 3 ']' + test_suite_name=ui + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/ui.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-ui make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true ./tests/e2e/ui/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.17.0-0.nightly-2024-09-29-173421 True False 16m Cluster version is 4.17.0-0.nightly-2024-09-29-173421' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.17.0-0.nightly-2024-09-29-173421 True False 16m Cluster version is 4.17.0-0.nightly-2024-09-29-173421' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/ui/render.sh ++ export SUITE_DIR=./tests/e2e/ui ++ SUITE_DIR=./tests/e2e/ui ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/ui ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test allinone + '[' 1 -ne 1 ']' + test_name=allinone + echo =========================================================================== =========================================================================== + info 'Rendering files for test allinone' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test allinone\e[0m' Rendering files for test allinone + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build + '[' _build '!=' _build ']' + mkdir -p allinone + cd allinone + export GET_URL_COMMAND + export URL + export JAEGER_NAME=all-in-one-ui + JAEGER_NAME=all-in-one-ui + '[' true = true ']' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./01-curl.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./04-test-ui-config.yaml + start_test production + '[' 1 -ne 1 ']' + test_name=production + echo =========================================================================== =========================================================================== + info 'Rendering files for test production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test production\e[0m' Rendering files for test production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build/allinone + '[' allinone '!=' _build ']' + cd .. + mkdir -p production + cd production + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + [[ true = true ]] + [[ true = true ]] + render_install_jaeger production-ui production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + '[' true = true ']' + INSECURE=true + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-forbbiden-access.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-curl.yaml + INSECURE=true + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./05-check-disabled-security.yaml + ASSERT_PRESENT=false + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./06-check-NO-gaID.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./08-check-gaID.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running ui E2E tests' Running ui E2E tests + cd tests/e2e/ui/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2856073267 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 3 tests === RUN kuttl/harness === RUN kuttl/harness/allinone === PAUSE kuttl/harness/allinone === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/production === PAUSE kuttl/harness/production === CONT kuttl/harness/allinone logger.go:42: 17:53:58 | allinone | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 17:53:58 | allinone | Creating namespace: kuttl-test-credible-longhorn logger.go:42: 17:53:58 | allinone/0-install | starting test step 0-install logger.go:42: 17:53:58 | allinone/0-install | Jaeger:kuttl-test-credible-longhorn/all-in-one-ui created logger.go:42: 17:54:02 | allinone/0-install | test step completed 0-install logger.go:42: 17:54:02 | allinone/1-curl | starting test step 1-curl logger.go:42: 17:54:02 | allinone/1-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 17:54:02 | allinone/1-curl | Checking the Ingress host value was populated logger.go:42: 17:54:02 | allinone/1-curl | Try number 0 logger.go:42: 17:54:02 | allinone/1-curl | Hostname is all-in-one-ui-kuttl-test-credible-longhorn.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com logger.go:42: 17:54:02 | allinone/1-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE all-in-one-ui] logger.go:42: 17:54:02 | allinone/1-curl | Checking an expected HTTP response logger.go:42: 17:54:02 | allinone/1-curl | Running in OpenShift logger.go:42: 17:54:02 | allinone/1-curl | User not provided. Getting the token... logger.go:42: 17:54:04 | allinone/1-curl | Warning: resource jaegers/all-in-one-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 17:54:13 | allinone/1-curl | Try number 1/30 the https://all-in-one-ui-kuttl-test-credible-longhorn.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com/search logger.go:42: 17:54:13 | allinone/1-curl | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 17:54:13 | allinone/1-curl | Try number 2/30 the https://all-in-one-ui-kuttl-test-credible-longhorn.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com/search logger.go:42: 17:54:13 | allinone/1-curl | HTTP response is 503. 200 expected. Waiting 10 s logger.go:42: 17:54:23 | allinone/1-curl | Try number 3/30 the https://all-in-one-ui-kuttl-test-credible-longhorn.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com/search logger.go:42: 17:54:23 | allinone/1-curl | curl response asserted properly logger.go:42: 17:54:23 | allinone/1-curl | test step completed 1-curl logger.go:42: 17:54:23 | allinone/2-delete | starting test step 2-delete logger.go:42: 17:54:23 | allinone/2-delete | Jaeger:kuttl-test-credible-longhorn/all-in-one-ui created logger.go:42: 17:54:23 | allinone/2-delete | test step completed 2-delete logger.go:42: 17:54:23 | allinone/3-install | starting test step 3-install logger.go:42: 17:54:23 | allinone/3-install | Jaeger:kuttl-test-credible-longhorn/all-in-one-ui updated logger.go:42: 17:54:23 | allinone/3-install | test step completed 3-install logger.go:42: 17:54:23 | allinone/4-test-ui-config | starting test step 4-test-ui-config logger.go:42: 17:54:23 | allinone/4-test-ui-config | running command: [./ensure-ingress-host.sh] logger.go:42: 17:54:23 | allinone/4-test-ui-config | Checking the Ingress host value was populated logger.go:42: 17:54:23 | allinone/4-test-ui-config | Try number 0 logger.go:42: 17:54:23 | allinone/4-test-ui-config | error: error executing jsonpath "{.items[0].status.ingress[0].host}": Error executing template: array index out of bounds: index 0, length 0. Printing more information for debugging the template: logger.go:42: 17:54:23 | allinone/4-test-ui-config | template was: logger.go:42: 17:54:23 | allinone/4-test-ui-config | {.items[0].status.ingress[0].host} logger.go:42: 17:54:23 | allinone/4-test-ui-config | object given to jsonpath engine was: logger.go:42: 17:54:23 | allinone/4-test-ui-config | map[string]interface {}{"apiVersion":"v1", "items":[]interface {}{}, "kind":"List", "metadata":map[string]interface {}{"resourceVersion":""}} logger.go:42: 17:54:23 | allinone/4-test-ui-config | logger.go:42: 17:54:23 | allinone/4-test-ui-config | logger.go:42: 17:54:33 | allinone/4-test-ui-config | Try number 1 logger.go:42: 17:54:33 | allinone/4-test-ui-config | Hostname is all-in-one-ui-kuttl-test-credible-longhorn.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com logger.go:42: 17:54:33 | allinone/4-test-ui-config | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 17:54:35 | allinone/4-test-ui-config | time="2024-09-30T17:54:35Z" level=info msg="Querying https://all-in-one-ui-kuttl-test-credible-longhorn.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com/search..." logger.go:42: 17:54:35 | allinone/4-test-ui-config | time="2024-09-30T17:54:35Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 17:54:35 | allinone/4-test-ui-config | time="2024-09-30T17:54:35Z" level=info msg="Polling to https://all-in-one-ui-kuttl-test-credible-longhorn.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com/search" logger.go:42: 17:54:35 | allinone/4-test-ui-config | time="2024-09-30T17:54:35Z" level=info msg="Doing request number 0" logger.go:42: 17:54:35 | allinone/4-test-ui-config | time="2024-09-30T17:54:35Z" level=info msg="Content found and asserted!" logger.go:42: 17:54:35 | allinone/4-test-ui-config | time="2024-09-30T17:54:35Z" level=info msg="Success!" logger.go:42: 17:54:35 | allinone/4-test-ui-config | test step completed 4-test-ui-config logger.go:42: 17:54:35 | allinone | allinone events from ns kuttl-test-credible-longhorn: logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:01 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-84464984 to 1 deployment-controller logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:02 +0000 UTC Normal Pod all-in-one-ui-84464984-zpqdp Binding Scheduled Successfully assigned kuttl-test-credible-longhorn/all-in-one-ui-84464984-zpqdp to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:02 +0000 UTC Normal Pod all-in-one-ui-84464984-zpqdp AddedInterface Add eth0 [10.131.0.35/23] from ovn-kubernetes multus logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:02 +0000 UTC Normal Pod all-in-one-ui-84464984-zpqdp.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:02 +0000 UTC Normal Pod all-in-one-ui-84464984-zpqdp.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:02 +0000 UTC Normal Pod all-in-one-ui-84464984-zpqdp.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:02 +0000 UTC Normal Pod all-in-one-ui-84464984-zpqdp.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:02 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-84464984 SuccessfulCreate Created pod: all-in-one-ui-84464984-zpqdp replicaset-controller logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:03 +0000 UTC Normal Pod all-in-one-ui-84464984-zpqdp.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:03 +0000 UTC Normal Pod all-in-one-ui-84464984-zpqdp.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:09 +0000 UTC Normal Pod all-in-one-ui-84464984-zpqdp.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:09 +0000 UTC Normal Pod all-in-one-ui-84464984-zpqdp.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:09 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-84464984 SuccessfulDelete Deleted pod: all-in-one-ui-84464984-zpqdp replicaset-controller logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:09 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled down replica set all-in-one-ui-84464984 to 0 from 1 deployment-controller logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:10 +0000 UTC Normal Pod all-in-one-ui-696b7d7d97-9n5n4 Binding Scheduled Successfully assigned kuttl-test-credible-longhorn/all-in-one-ui-696b7d7d97-9n5n4 to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:10 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-696b7d7d97 SuccessfulCreate Created pod: all-in-one-ui-696b7d7d97-9n5n4 replicaset-controller logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:10 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-696b7d7d97 to 1 deployment-controller logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:11 +0000 UTC Normal Pod all-in-one-ui-696b7d7d97-9n5n4 AddedInterface Add eth0 [10.131.0.36/23] from ovn-kubernetes multus logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:11 +0000 UTC Normal Pod all-in-one-ui-696b7d7d97-9n5n4.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:11 +0000 UTC Normal Pod all-in-one-ui-696b7d7d97-9n5n4.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:11 +0000 UTC Normal Pod all-in-one-ui-696b7d7d97-9n5n4.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:11 +0000 UTC Normal Pod all-in-one-ui-696b7d7d97-9n5n4.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:11 +0000 UTC Normal Pod all-in-one-ui-696b7d7d97-9n5n4.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:11 +0000 UTC Normal Pod all-in-one-ui-696b7d7d97-9n5n4.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:23 +0000 UTC Normal Pod all-in-one-ui-696b7d7d97-9n5n4.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:23 +0000 UTC Normal Pod all-in-one-ui-696b7d7d97-9n5n4.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:29 +0000 UTC Normal Pod all-in-one-ui-f89c594bd-8gxrf Binding Scheduled Successfully assigned kuttl-test-credible-longhorn/all-in-one-ui-f89c594bd-8gxrf to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:29 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-f89c594bd SuccessfulCreate Created pod: all-in-one-ui-f89c594bd-8gxrf replicaset-controller logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:29 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-f89c594bd to 1 deployment-controller logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:30 +0000 UTC Normal Pod all-in-one-ui-f89c594bd-8gxrf AddedInterface Add eth0 [10.131.0.37/23] from ovn-kubernetes multus logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:30 +0000 UTC Normal Pod all-in-one-ui-f89c594bd-8gxrf.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:30 +0000 UTC Normal Pod all-in-one-ui-f89c594bd-8gxrf.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 17:54:35 | allinone | 2024-09-30 17:54:30 +0000 UTC Normal Pod all-in-one-ui-f89c594bd-8gxrf.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 17:54:35 | allinone | Deleting namespace: kuttl-test-credible-longhorn === CONT kuttl/harness/production logger.go:42: 17:54:42 | production | Ignoring add-tracking-id.yaml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 17:54:42 | production | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 17:54:42 | production | Creating namespace: kuttl-test-warm-coral logger.go:42: 17:54:42 | production/1-install | starting test step 1-install logger.go:42: 17:54:42 | production/1-install | Jaeger:kuttl-test-warm-coral/production-ui created logger.go:42: 17:55:26 | production/1-install | test step completed 1-install logger.go:42: 17:55:26 | production/2-check-forbbiden-access | starting test step 2-check-forbbiden-access logger.go:42: 17:55:26 | production/2-check-forbbiden-access | running command: [./ensure-ingress-host.sh] logger.go:42: 17:55:26 | production/2-check-forbbiden-access | Checking the Ingress host value was populated logger.go:42: 17:55:26 | production/2-check-forbbiden-access | Try number 0 logger.go:42: 17:55:26 | production/2-check-forbbiden-access | Hostname is production-ui-kuttl-test-warm-coral.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com logger.go:42: 17:55:26 | production/2-check-forbbiden-access | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE production-ui] logger.go:42: 17:55:27 | production/2-check-forbbiden-access | Checking an expected HTTP response logger.go:42: 17:55:27 | production/2-check-forbbiden-access | Running in OpenShift logger.go:42: 17:55:27 | production/2-check-forbbiden-access | Not using any secret logger.go:42: 17:55:27 | production/2-check-forbbiden-access | Try number 1/30 the https://production-ui-kuttl-test-warm-coral.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com/search logger.go:42: 17:55:27 | production/2-check-forbbiden-access | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 17:55:27 | production/2-check-forbbiden-access | Try number 2/30 the https://production-ui-kuttl-test-warm-coral.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com/search logger.go:42: 17:55:27 | production/2-check-forbbiden-access | curl response asserted properly logger.go:42: 17:55:27 | production/2-check-forbbiden-access | test step completed 2-check-forbbiden-access logger.go:42: 17:55:27 | production/3-curl | starting test step 3-curl logger.go:42: 17:55:27 | production/3-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 17:55:27 | production/3-curl | Checking the Ingress host value was populated logger.go:42: 17:55:27 | production/3-curl | Try number 0 logger.go:42: 17:55:27 | production/3-curl | Hostname is production-ui-kuttl-test-warm-coral.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com logger.go:42: 17:55:27 | production/3-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 17:55:27 | production/3-curl | Checking an expected HTTP response logger.go:42: 17:55:27 | production/3-curl | Running in OpenShift logger.go:42: 17:55:27 | production/3-curl | User not provided. Getting the token... logger.go:42: 17:55:29 | production/3-curl | Warning: resource jaegers/production-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 17:55:35 | production/3-curl | Try number 1/30 the https://production-ui-kuttl-test-warm-coral.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com/search logger.go:42: 17:55:35 | production/3-curl | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 17:55:35 | production/3-curl | Try number 2/30 the https://production-ui-kuttl-test-warm-coral.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com/search logger.go:42: 17:55:36 | production/3-curl | curl response asserted properly logger.go:42: 17:55:36 | production/3-curl | test step completed 3-curl logger.go:42: 17:55:36 | production/4-install | starting test step 4-install logger.go:42: 17:55:36 | production/4-install | Jaeger:kuttl-test-warm-coral/production-ui updated logger.go:42: 17:55:36 | production/4-install | test step completed 4-install logger.go:42: 17:55:36 | production/5-check-disabled-security | starting test step 5-check-disabled-security logger.go:42: 17:55:36 | production/5-check-disabled-security | running command: [./ensure-ingress-host.sh] logger.go:42: 17:55:36 | production/5-check-disabled-security | Checking the Ingress host value was populated logger.go:42: 17:55:36 | production/5-check-disabled-security | Try number 0 logger.go:42: 17:55:36 | production/5-check-disabled-security | Hostname is production-ui-kuttl-test-warm-coral.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com logger.go:42: 17:55:36 | production/5-check-disabled-security | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 17:55:36 | production/5-check-disabled-security | Checking an expected HTTP response logger.go:42: 17:55:36 | production/5-check-disabled-security | Running in OpenShift logger.go:42: 17:55:36 | production/5-check-disabled-security | Not using any secret logger.go:42: 17:55:36 | production/5-check-disabled-security | Try number 1/30 the https://production-ui-kuttl-test-warm-coral.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com/search logger.go:42: 17:55:36 | production/5-check-disabled-security | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 17:55:36 | production/5-check-disabled-security | Try number 2/30 the https://production-ui-kuttl-test-warm-coral.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com/search logger.go:42: 17:55:36 | production/5-check-disabled-security | HTTP response is 403. 200 expected. Waiting 10 s logger.go:42: 17:55:46 | production/5-check-disabled-security | Try number 3/30 the https://production-ui-kuttl-test-warm-coral.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com/search logger.go:42: 17:55:46 | production/5-check-disabled-security | curl response asserted properly logger.go:42: 17:55:46 | production/5-check-disabled-security | test step completed 5-check-disabled-security logger.go:42: 17:55:46 | production/6-check-NO-gaID | starting test step 6-check-NO-gaID logger.go:42: 17:55:46 | production/6-check-NO-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 17:55:46 | production/6-check-NO-gaID | Checking the Ingress host value was populated logger.go:42: 17:55:46 | production/6-check-NO-gaID | Try number 0 logger.go:42: 17:55:46 | production/6-check-NO-gaID | Hostname is production-ui-kuttl-test-warm-coral.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com logger.go:42: 17:55:46 | production/6-check-NO-gaID | running command: [sh -c ASSERT_PRESENT=false EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 17:55:47 | production/6-check-NO-gaID | time="2024-09-30T17:55:47Z" level=info msg="Querying https://production-ui-kuttl-test-warm-coral.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com/search..." logger.go:42: 17:55:47 | production/6-check-NO-gaID | time="2024-09-30T17:55:47Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 17:55:47 | production/6-check-NO-gaID | time="2024-09-30T17:55:47Z" level=info msg="Polling to https://production-ui-kuttl-test-warm-coral.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com/search" logger.go:42: 17:55:47 | production/6-check-NO-gaID | time="2024-09-30T17:55:47Z" level=info msg="Doing request number 0" logger.go:42: 17:55:47 | production/6-check-NO-gaID | time="2024-09-30T17:55:47Z" level=info msg="Content not found and asserted it was not found!" logger.go:42: 17:55:47 | production/6-check-NO-gaID | time="2024-09-30T17:55:47Z" level=info msg="Success!" logger.go:42: 17:55:47 | production/6-check-NO-gaID | test step completed 6-check-NO-gaID logger.go:42: 17:55:47 | production/7-add-tracking-id | starting test step 7-add-tracking-id logger.go:42: 17:55:47 | production/7-add-tracking-id | running command: [sh -c kubectl apply -f add-tracking-id.yaml -n $NAMESPACE] logger.go:42: 17:55:49 | production/7-add-tracking-id | jaeger.jaegertracing.io/production-ui configured logger.go:42: 17:55:49 | production/7-add-tracking-id | test step completed 7-add-tracking-id logger.go:42: 17:55:49 | production/8-check-gaID | starting test step 8-check-gaID logger.go:42: 17:55:49 | production/8-check-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 17:55:49 | production/8-check-gaID | Checking the Ingress host value was populated logger.go:42: 17:55:49 | production/8-check-gaID | Try number 0 logger.go:42: 17:55:49 | production/8-check-gaID | Hostname is production-ui-kuttl-test-warm-coral.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com logger.go:42: 17:55:49 | production/8-check-gaID | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 17:55:50 | production/8-check-gaID | time="2024-09-30T17:55:50Z" level=info msg="Querying https://production-ui-kuttl-test-warm-coral.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com/search..." logger.go:42: 17:55:50 | production/8-check-gaID | time="2024-09-30T17:55:50Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 17:55:50 | production/8-check-gaID | time="2024-09-30T17:55:50Z" level=info msg="Polling to https://production-ui-kuttl-test-warm-coral.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com/search" logger.go:42: 17:55:50 | production/8-check-gaID | time="2024-09-30T17:55:50Z" level=info msg="Doing request number 0" logger.go:42: 17:55:50 | production/8-check-gaID | time="2024-09-30T17:55:50Z" level=warning msg="Found: false . Assert: true" logger.go:42: 17:55:50 | production/8-check-gaID | time="2024-09-30T17:55:50Z" level=warning msg="The condition of the test function was not accomplished" logger.go:42: 17:55:58 | production/8-check-gaID | time="2024-09-30T17:55:58Z" level=info msg="Doing request number 1" logger.go:42: 17:56:18 | production/8-check-gaID | time="2024-09-30T17:56:18Z" level=info msg="Content found and asserted!" logger.go:42: 17:56:18 | production/8-check-gaID | time="2024-09-30T17:56:18Z" level=info msg="Success!" logger.go:42: 17:56:18 | production/8-check-gaID | test step completed 8-check-gaID logger.go:42: 17:56:18 | production | production events from ns kuttl-test-warm-coral: logger.go:42: 17:56:18 | production | 2024-09-30 17:54:49 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestwarmcoralproductionui-1-5ffc6c9bb8 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestwarmcoralproductionui-1-5ffc6c9ggtfk replicaset-controller logger.go:42: 17:56:18 | production | 2024-09-30 17:54:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwarmcoralproductionui-1-5ffc6c9ggtfk Binding Scheduled Successfully assigned kuttl-test-warm-coral/elasticsearch-cdm-kuttltestwarmcoralproductionui-1-5ffc6c9ggtfk to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 17:56:18 | production | 2024-09-30 17:54:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwarmcoralproductionui-1-5ffc6c9ggtfk AddedInterface Add eth0 [10.128.2.22/23] from ovn-kubernetes multus logger.go:42: 17:56:18 | production | 2024-09-30 17:54:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwarmcoralproductionui-1-5ffc6c9ggtfk.spec.containers{elasticsearch} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:89629964e44058bafc90393a235b08c7c974f05513aecbfe7229134da732f2b5" kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:54:49 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestwarmcoralproductionui-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestwarmcoralproductionui-1-5ffc6c9bb8 to 1 deployment-controller logger.go:42: 17:56:18 | production | 2024-09-30 17:54:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwarmcoralproductionui-1-5ffc6c9ggtfk.spec.containers{elasticsearch} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:89629964e44058bafc90393a235b08c7c974f05513aecbfe7229134da732f2b5" in 4.727s (4.727s including waiting). Image size: 523661080 bytes. kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:54:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwarmcoralproductionui-1-5ffc6c9ggtfk.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:54:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwarmcoralproductionui-1-5ffc6c9ggtfk.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:54:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwarmcoralproductionui-1-5ffc6c9ggtfk.spec.containers{proxy} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:a8f53dec42a46c5bf8ac7f2888848c01e70f405d27211e2a269730c888929faf" kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:54:57 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwarmcoralproductionui-1-5ffc6c9ggtfk.spec.containers{proxy} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:a8f53dec42a46c5bf8ac7f2888848c01e70f405d27211e2a269730c888929faf" in 2.83s (2.83s including waiting). Image size: 272142633 bytes. kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:54:57 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwarmcoralproductionui-1-5ffc6c9ggtfk.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:54:57 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwarmcoralproductionui-1-5ffc6c9ggtfk.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:04 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestwarmcoralproductionui-1-5ffc6c9ggtfk.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:09 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestwarmcoralproductionui-1-5ffc6c9ggtfk.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:20 +0000 UTC Normal Pod production-ui-collector-b4dff7d8b-zkbwn Binding Scheduled Successfully assigned kuttl-test-warm-coral/production-ui-collector-b4dff7d8b-zkbwn to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 17:56:18 | production | 2024-09-30 17:55:20 +0000 UTC Normal Pod production-ui-collector-b4dff7d8b-zkbwn AddedInterface Add eth0 [10.129.2.29/23] from ovn-kubernetes multus logger.go:42: 17:56:18 | production | 2024-09-30 17:55:20 +0000 UTC Normal Pod production-ui-collector-b4dff7d8b-zkbwn.spec.containers{jaeger-collector} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:20 +0000 UTC Normal ReplicaSet.apps production-ui-collector-b4dff7d8b SuccessfulCreate Created pod: production-ui-collector-b4dff7d8b-zkbwn replicaset-controller logger.go:42: 17:56:18 | production | 2024-09-30 17:55:20 +0000 UTC Normal Deployment.apps production-ui-collector ScalingReplicaSet Scaled up replica set production-ui-collector-b4dff7d8b to 1 deployment-controller logger.go:42: 17:56:18 | production | 2024-09-30 17:55:20 +0000 UTC Normal Pod production-ui-query-79cc5498f6-skr7q Binding Scheduled Successfully assigned kuttl-test-warm-coral/production-ui-query-79cc5498f6-skr7q to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 17:56:18 | production | 2024-09-30 17:55:20 +0000 UTC Normal Pod production-ui-query-79cc5498f6-skr7q AddedInterface Add eth0 [10.131.0.38/23] from ovn-kubernetes multus logger.go:42: 17:56:18 | production | 2024-09-30 17:55:20 +0000 UTC Normal ReplicaSet.apps production-ui-query-79cc5498f6 SuccessfulCreate Created pod: production-ui-query-79cc5498f6-skr7q replicaset-controller logger.go:42: 17:56:18 | production | 2024-09-30 17:55:20 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-79cc5498f6 to 1 deployment-controller logger.go:42: 17:56:18 | production | 2024-09-30 17:55:21 +0000 UTC Normal Pod production-ui-query-79cc5498f6-skr7q.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:21 +0000 UTC Normal Pod production-ui-query-79cc5498f6-skr7q.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:21 +0000 UTC Normal Pod production-ui-query-79cc5498f6-skr7q.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:21 +0000 UTC Normal Pod production-ui-query-79cc5498f6-skr7q.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:21 +0000 UTC Normal Pod production-ui-query-79cc5498f6-skr7q.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:21 +0000 UTC Normal Pod production-ui-query-79cc5498f6-skr7q.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:21 +0000 UTC Normal Pod production-ui-query-79cc5498f6-skr7q.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:21 +0000 UTC Normal Pod production-ui-query-79cc5498f6-skr7q.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:21 +0000 UTC Normal Pod production-ui-query-79cc5498f6-skr7q.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:24 +0000 UTC Normal Pod production-ui-collector-b4dff7d8b-zkbwn.spec.containers{jaeger-collector} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" in 3.954s (3.954s including waiting). Image size: 142315153 bytes. kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:25 +0000 UTC Normal Pod production-ui-collector-b4dff7d8b-zkbwn.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:25 +0000 UTC Normal Pod production-ui-collector-b4dff7d8b-zkbwn.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:30 +0000 UTC Normal Pod production-ui-query-79cc5498f6-skr7q.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:30 +0000 UTC Normal Pod production-ui-query-79cc5498f6-skr7q.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:30 +0000 UTC Normal Pod production-ui-query-79cc5498f6-skr7q.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:30 +0000 UTC Normal ReplicaSet.apps production-ui-query-79cc5498f6 SuccessfulDelete Deleted pod: production-ui-query-79cc5498f6-skr7q replicaset-controller logger.go:42: 17:56:18 | production | 2024-09-30 17:55:30 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-79cc5498f6 to 0 from 1 deployment-controller logger.go:42: 17:56:18 | production | 2024-09-30 17:55:31 +0000 UTC Normal Pod production-ui-query-8b59dd889-w4bqm Binding Scheduled Successfully assigned kuttl-test-warm-coral/production-ui-query-8b59dd889-w4bqm to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 17:56:18 | production | 2024-09-30 17:55:31 +0000 UTC Normal ReplicaSet.apps production-ui-query-8b59dd889 SuccessfulCreate Created pod: production-ui-query-8b59dd889-w4bqm replicaset-controller logger.go:42: 17:56:18 | production | 2024-09-30 17:55:31 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-8b59dd889 to 1 deployment-controller logger.go:42: 17:56:18 | production | 2024-09-30 17:55:32 +0000 UTC Normal Pod production-ui-query-8b59dd889-w4bqm AddedInterface Add eth0 [10.131.0.39/23] from ovn-kubernetes multus logger.go:42: 17:56:18 | production | 2024-09-30 17:55:32 +0000 UTC Normal Pod production-ui-query-8b59dd889-w4bqm.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:32 +0000 UTC Normal Pod production-ui-query-8b59dd889-w4bqm.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:32 +0000 UTC Normal Pod production-ui-query-8b59dd889-w4bqm.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:32 +0000 UTC Normal Pod production-ui-query-8b59dd889-w4bqm.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:32 +0000 UTC Normal Pod production-ui-query-8b59dd889-w4bqm.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:32 +0000 UTC Normal Pod production-ui-query-8b59dd889-w4bqm.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:32 +0000 UTC Normal Pod production-ui-query-8b59dd889-w4bqm.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:32 +0000 UTC Normal Pod production-ui-query-8b59dd889-w4bqm.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:32 +0000 UTC Normal Pod production-ui-query-8b59dd889-w4bqm.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:35 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 17:56:18 | production | 2024-09-30 17:55:35 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 17:56:18 | production | 2024-09-30 17:55:35 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 17:56:18 | production | 2024-09-30 17:55:40 +0000 UTC Normal Pod production-ui-query-8b59dd889-w4bqm.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:40 +0000 UTC Normal Pod production-ui-query-8b59dd889-w4bqm.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:40 +0000 UTC Normal Pod production-ui-query-8b59dd889-w4bqm.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:40 +0000 UTC Normal ReplicaSet.apps production-ui-query-8b59dd889 SuccessfulDelete Deleted pod: production-ui-query-8b59dd889-w4bqm replicaset-controller logger.go:42: 17:56:18 | production | 2024-09-30 17:55:40 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-8b59dd889 to 0 from 1 deployment-controller logger.go:42: 17:56:18 | production | 2024-09-30 17:55:41 +0000 UTC Normal Pod production-ui-query-6f7f48764b-2rhhp Binding Scheduled Successfully assigned kuttl-test-warm-coral/production-ui-query-6f7f48764b-2rhhp to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 17:56:18 | production | 2024-09-30 17:55:41 +0000 UTC Normal ReplicaSet.apps production-ui-query-6f7f48764b SuccessfulCreate Created pod: production-ui-query-6f7f48764b-2rhhp replicaset-controller logger.go:42: 17:56:18 | production | 2024-09-30 17:55:41 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-6f7f48764b to 1 deployment-controller logger.go:42: 17:56:18 | production | 2024-09-30 17:55:42 +0000 UTC Normal Pod production-ui-query-6f7f48764b-2rhhp AddedInterface Add eth0 [10.131.0.40/23] from ovn-kubernetes multus logger.go:42: 17:56:18 | production | 2024-09-30 17:55:42 +0000 UTC Normal Pod production-ui-query-6f7f48764b-2rhhp.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:42 +0000 UTC Normal Pod production-ui-query-6f7f48764b-2rhhp.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:42 +0000 UTC Normal Pod production-ui-query-6f7f48764b-2rhhp.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:42 +0000 UTC Normal Pod production-ui-query-6f7f48764b-2rhhp.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:42 +0000 UTC Normal Pod production-ui-query-6f7f48764b-2rhhp.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:42 +0000 UTC Normal Pod production-ui-query-6f7f48764b-2rhhp.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:50 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 17:56:18 | production | 2024-09-30 17:55:50 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod production-ui-collector-b4dff7d8b-zkbwn horizontal-pod-autoscaler logger.go:42: 17:56:18 | production | 2024-09-30 17:55:50 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 17:56:18 | production | 2024-09-30 17:55:50 +0000 UTC Normal Pod production-ui-query-6f7f48764b-2rhhp.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:50 +0000 UTC Normal Pod production-ui-query-6f7f48764b-2rhhp.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:50 +0000 UTC Normal ReplicaSet.apps production-ui-query-6f7f48764b SuccessfulDelete Deleted pod: production-ui-query-6f7f48764b-2rhhp replicaset-controller logger.go:42: 17:56:18 | production | 2024-09-30 17:55:50 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-6f7f48764b to 0 from 1 deployment-controller logger.go:42: 17:56:18 | production | 2024-09-30 17:55:51 +0000 UTC Normal Pod production-ui-query-6675fc8bbf-wj6ft Binding Scheduled Successfully assigned kuttl-test-warm-coral/production-ui-query-6675fc8bbf-wj6ft to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 17:56:18 | production | 2024-09-30 17:55:51 +0000 UTC Normal ReplicaSet.apps production-ui-query-6675fc8bbf SuccessfulCreate Created pod: production-ui-query-6675fc8bbf-wj6ft replicaset-controller logger.go:42: 17:56:18 | production | 2024-09-30 17:55:51 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-6675fc8bbf to 1 deployment-controller logger.go:42: 17:56:18 | production | 2024-09-30 17:55:52 +0000 UTC Normal Pod production-ui-query-6675fc8bbf-wj6ft AddedInterface Add eth0 [10.131.0.41/23] from ovn-kubernetes multus logger.go:42: 17:56:18 | production | 2024-09-30 17:55:52 +0000 UTC Normal Pod production-ui-query-6675fc8bbf-wj6ft.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:52 +0000 UTC Normal Pod production-ui-query-6675fc8bbf-wj6ft.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:52 +0000 UTC Normal Pod production-ui-query-6675fc8bbf-wj6ft.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:52 +0000 UTC Normal Pod production-ui-query-6675fc8bbf-wj6ft.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:52 +0000 UTC Normal Pod production-ui-query-6675fc8bbf-wj6ft.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:55:52 +0000 UTC Normal Pod production-ui-query-6675fc8bbf-wj6ft.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 17:56:18 | production | 2024-09-30 17:56:05 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod production-ui-collector-b4dff7d8b-zkbwn horizontal-pod-autoscaler logger.go:42: 17:56:18 | production | 2024-09-30 17:56:05 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod production-ui-collector-b4dff7d8b-zkbwn horizontal-pod-autoscaler logger.go:42: 17:56:18 | production | Deleting namespace: kuttl-test-warm-coral === CONT kuttl/harness/artifacts logger.go:42: 17:56:25 | artifacts | Creating namespace: kuttl-test-innocent-ox logger.go:42: 17:56:25 | artifacts | artifacts events from ns kuttl-test-innocent-ox: logger.go:42: 17:56:25 | artifacts | Deleting namespace: kuttl-test-innocent-ox === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (153.24s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/allinone (44.35s) --- PASS: kuttl/harness/production (102.81s) --- PASS: kuttl/harness/artifacts (5.89s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name ui --report --output /logs/artifacts/ui.xml ./artifacts/kuttl-report.xml time="2024-09-30T17:56:31Z" level=debug msg="Setting a new name for the test suites" time="2024-09-30T17:56:31Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-09-30T17:56:31Z" level=debug msg="normalizing test case names" time="2024-09-30T17:56:31Z" level=debug msg="ui/allinone -> ui_allinone" time="2024-09-30T17:56:31Z" level=debug msg="ui/production -> ui_production" time="2024-09-30T17:56:31Z" level=debug msg="ui/artifacts -> ui_artifacts" +---------------+--------+ | NAME | RESULT | +---------------+--------+ | ui_allinone | passed | | ui_production | passed | | ui_artifacts | passed | +---------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh miscellaneous false true + '[' 3 -ne 3 ']' + test_suite_name=miscellaneous + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/miscellaneous.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-miscellaneous make[2]: Entering directory '/tmp/jaeger-tests' SKIP_ES_EXTERNAL=true ./tests/e2e/miscellaneous/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.17.0-0.nightly-2024-09-29-173421 True False 19m Cluster version is 4.17.0-0.nightly-2024-09-29-173421' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.17.0-0.nightly-2024-09-29-173421 True False 19m Cluster version is 4.17.0-0.nightly-2024-09-29-173421' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/miscellaneous/render.sh ++ export SUITE_DIR=./tests/e2e/miscellaneous ++ SUITE_DIR=./tests/e2e/miscellaneous ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/miscellaneous ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test collector-autoscale + '[' 1 -ne 1 ']' + test_name=collector-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-autoscale\e[0m' Rendering files for test collector-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p collector-autoscale + cd collector-autoscale + jaeger_name=simple-prod + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + ELASTICSEARCH_NODECOUNT=1 + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.resources.requests.memory="200m"' 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.autoscale=true 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.minReplicas=1 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.maxReplicas=2 01-install.yaml + version_lt 1.30 1.23 ++ sort -rV ++ tr ' ' '\n' ++ head -n 1 ++ echo 1.30 1.23 + test 1.30 '!=' 1.30 + rm ./03-assert.yaml + generate_otlp_e2e_tests http + test_protocol=http + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-http\e[0m' Rendering files for test collector-otlp-allinone-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-autoscale + '[' collector-autoscale '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-http + cd collector-otlp-allinone-http + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger http true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-http\e[0m' Rendering files for test collector-otlp-production-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-http + '[' collector-otlp-allinone-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-http + cd collector-otlp-production-http + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger http true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + generate_otlp_e2e_tests grpc + test_protocol=grpc + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-grpc\e[0m' Rendering files for test collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-http + '[' collector-otlp-production-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-grpc + cd collector-otlp-allinone-grpc + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-grpc\e[0m' Rendering files for test collector-otlp-production-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-grpc + '[' collector-otlp-allinone-grpc '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-grpc + cd collector-otlp-production-grpc + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + '[' true = true ']' + skip_test istio 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=istio + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-grpc + '[' collector-otlp-production-grpc '!=' _build ']' + cd .. + rm -rf istio + warning 'istio: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: istio: Test not supported in OpenShift\e[0m' WAR: istio: Test not supported in OpenShift + '[' true = true ']' + skip_test outside-cluster 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=outside-cluster + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + rm -rf outside-cluster + warning 'outside-cluster: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: outside-cluster: Test not supported in OpenShift\e[0m' WAR: outside-cluster: Test not supported in OpenShift + start_test set-custom-img + '[' 1 -ne 1 ']' + test_name=set-custom-img + echo =========================================================================== =========================================================================== + info 'Rendering files for test set-custom-img' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test set-custom-img\e[0m' Rendering files for test set-custom-img + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p set-custom-img + cd set-custom-img + jaeger_name=my-jaeger + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.image="test"' ./02-install.yaml + '[' true = true ']' + skip_test non-cluster-wide 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=non-cluster-wide + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/set-custom-img + '[' set-custom-img '!=' _build ']' + cd .. + rm -rf non-cluster-wide + warning 'non-cluster-wide: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: non-cluster-wide: Test not supported in OpenShift\e[0m' WAR: non-cluster-wide: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running miscellaneous E2E tests' Running miscellaneous E2E tests + cd tests/e2e/miscellaneous/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2856073267 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 8 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/cassandra-spark === PAUSE kuttl/harness/cassandra-spark === RUN kuttl/harness/collector-autoscale === PAUSE kuttl/harness/collector-autoscale === RUN kuttl/harness/collector-otlp-allinone-grpc === PAUSE kuttl/harness/collector-otlp-allinone-grpc === RUN kuttl/harness/collector-otlp-allinone-http === PAUSE kuttl/harness/collector-otlp-allinone-http === RUN kuttl/harness/collector-otlp-production-grpc === PAUSE kuttl/harness/collector-otlp-production-grpc === RUN kuttl/harness/collector-otlp-production-http === PAUSE kuttl/harness/collector-otlp-production-http === RUN kuttl/harness/set-custom-img === PAUSE kuttl/harness/set-custom-img === CONT kuttl/harness/artifacts logger.go:42: 17:56:43 | artifacts | Creating namespace: kuttl-test-cosmic-mackerel logger.go:42: 17:56:43 | artifacts | artifacts events from ns kuttl-test-cosmic-mackerel: logger.go:42: 17:56:43 | artifacts | Deleting namespace: kuttl-test-cosmic-mackerel === CONT kuttl/harness/collector-otlp-allinone-http logger.go:42: 17:56:50 | collector-otlp-allinone-http | Creating namespace: kuttl-test-wired-bluebird logger.go:42: 17:56:50 | collector-otlp-allinone-http/0-install | starting test step 0-install logger.go:42: 17:56:50 | collector-otlp-allinone-http/0-install | Jaeger:kuttl-test-wired-bluebird/my-jaeger created logger.go:42: 17:56:56 | collector-otlp-allinone-http/0-install | test step completed 0-install logger.go:42: 17:56:56 | collector-otlp-allinone-http/1-smoke-test | starting test step 1-smoke-test logger.go:42: 17:56:56 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 17:56:57 | collector-otlp-allinone-http/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 17:57:04 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 17:57:05 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 17:57:05 | collector-otlp-allinone-http/1-smoke-test | job.batch/report-span created logger.go:42: 17:57:05 | collector-otlp-allinone-http/1-smoke-test | job.batch/check-span created logger.go:42: 17:57:17 | collector-otlp-allinone-http/1-smoke-test | test step completed 1-smoke-test logger.go:42: 17:57:17 | collector-otlp-allinone-http | collector-otlp-allinone-http events from ns kuttl-test-wired-bluebird: logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:56:53 +0000 UTC Normal ReplicaSet.apps my-jaeger-74b9666654 SuccessfulCreate Created pod: my-jaeger-74b9666654-wx2bs replicaset-controller logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:56:53 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-74b9666654 to 1 deployment-controller logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:56:54 +0000 UTC Normal Pod my-jaeger-74b9666654-wx2bs Binding Scheduled Successfully assigned kuttl-test-wired-bluebird/my-jaeger-74b9666654-wx2bs to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:56:54 +0000 UTC Normal Pod my-jaeger-74b9666654-wx2bs AddedInterface Add eth0 [10.131.0.42/23] from ovn-kubernetes multus logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:56:54 +0000 UTC Normal Pod my-jaeger-74b9666654-wx2bs.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:56:54 +0000 UTC Normal Pod my-jaeger-74b9666654-wx2bs.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:56:54 +0000 UTC Normal Pod my-jaeger-74b9666654-wx2bs.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:56:54 +0000 UTC Normal Pod my-jaeger-74b9666654-wx2bs.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:56:54 +0000 UTC Normal Pod my-jaeger-74b9666654-wx2bs.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:56:54 +0000 UTC Normal Pod my-jaeger-74b9666654-wx2bs.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:56:59 +0000 UTC Normal Pod my-jaeger-74b9666654-wx2bs.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:56:59 +0000 UTC Normal Pod my-jaeger-74b9666654-wx2bs.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:56:59 +0000 UTC Normal ReplicaSet.apps my-jaeger-74b9666654 SuccessfulDelete Deleted pod: my-jaeger-74b9666654-wx2bs replicaset-controller logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:56:59 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-74b9666654 to 0 from 1 deployment-controller logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:57:00 +0000 UTC Normal Pod my-jaeger-d958fc646-zj2k2 Binding Scheduled Successfully assigned kuttl-test-wired-bluebird/my-jaeger-d958fc646-zj2k2 to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:57:00 +0000 UTC Normal ReplicaSet.apps my-jaeger-d958fc646 SuccessfulCreate Created pod: my-jaeger-d958fc646-zj2k2 replicaset-controller logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:57:00 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-d958fc646 to 1 deployment-controller logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:57:01 +0000 UTC Normal Pod my-jaeger-d958fc646-zj2k2 AddedInterface Add eth0 [10.131.0.43/23] from ovn-kubernetes multus logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:57:01 +0000 UTC Normal Pod my-jaeger-d958fc646-zj2k2.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:57:01 +0000 UTC Normal Pod my-jaeger-d958fc646-zj2k2.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:57:01 +0000 UTC Normal Pod my-jaeger-d958fc646-zj2k2.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:57:01 +0000 UTC Normal Pod my-jaeger-d958fc646-zj2k2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:57:01 +0000 UTC Normal Pod my-jaeger-d958fc646-zj2k2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:57:01 +0000 UTC Normal Pod my-jaeger-d958fc646-zj2k2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:57:05 +0000 UTC Normal Pod check-span-4h84h Binding Scheduled Successfully assigned kuttl-test-wired-bluebird/check-span-4h84h to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:57:05 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-4h84h job-controller logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:57:05 +0000 UTC Normal Pod report-span-dq7pm Binding Scheduled Successfully assigned kuttl-test-wired-bluebird/report-span-dq7pm to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:57:05 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-dq7pm job-controller logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:57:06 +0000 UTC Normal Pod check-span-4h84h AddedInterface Add eth0 [10.128.2.23/23] from ovn-kubernetes multus logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:57:06 +0000 UTC Normal Pod check-span-4h84h.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:57:06 +0000 UTC Normal Pod check-span-4h84h.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 212ms (212ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:57:06 +0000 UTC Normal Pod check-span-4h84h.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:57:06 +0000 UTC Normal Pod check-span-4h84h.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:57:06 +0000 UTC Normal Pod report-span-dq7pm AddedInterface Add eth0 [10.129.2.30/23] from ovn-kubernetes multus logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:57:06 +0000 UTC Normal Pod report-span-dq7pm.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:57:06 +0000 UTC Normal Pod report-span-dq7pm.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 273ms (273ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:57:06 +0000 UTC Normal Pod report-span-dq7pm.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:57:06 +0000 UTC Normal Pod report-span-dq7pm.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 17:57:17 | collector-otlp-allinone-http | 2024-09-30 17:57:17 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 17:57:17 | collector-otlp-allinone-http | Deleting namespace: kuttl-test-wired-bluebird === CONT kuttl/harness/set-custom-img logger.go:42: 17:57:29 | set-custom-img | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 17:57:29 | set-custom-img | Ignoring check-collector-img.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 17:57:29 | set-custom-img | Creating namespace: kuttl-test-apparent-man logger.go:42: 17:57:29 | set-custom-img/1-install | starting test step 1-install logger.go:42: 17:57:29 | set-custom-img/1-install | Jaeger:kuttl-test-apparent-man/my-jaeger created logger.go:42: 17:58:08 | set-custom-img/1-install | test step completed 1-install logger.go:42: 17:58:08 | set-custom-img/2-install | starting test step 2-install logger.go:42: 17:58:08 | set-custom-img/2-install | Jaeger:kuttl-test-apparent-man/my-jaeger updated logger.go:42: 17:58:08 | set-custom-img/2-install | test step completed 2-install logger.go:42: 17:58:08 | set-custom-img/3-check-image | starting test step 3-check-image logger.go:42: 17:58:08 | set-custom-img/3-check-image | running command: [sh -c ./check-collector-img.sh] logger.go:42: 17:58:08 | set-custom-img/3-check-image | Collector image missmatch. Expected: test. Has: registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c logger.go:42: 17:58:13 | set-custom-img/3-check-image | Collector image asserted properly! logger.go:42: 17:58:13 | set-custom-img/3-check-image | test step completed 3-check-image logger.go:42: 17:58:13 | set-custom-img | set-custom-img events from ns kuttl-test-apparent-man: logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:57:36 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestapparentmanmyjaeger-1-7dc744766 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestapparentmanmyjaeger-1-7dc744766hfj77 replicaset-controller logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:57:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestapparentmanmyjaeger-1-7dc744766hfj77 Binding Scheduled Successfully assigned kuttl-test-apparent-man/elasticsearch-cdm-kuttltestapparentmanmyjaeger-1-7dc744766hfj77 to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:57:36 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestapparentmanmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestapparentmanmyjaeger-1-7dc744766 to 1 deployment-controller logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:57:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestapparentmanmyjaeger-1-7dc744766hfj77 AddedInterface Add eth0 [10.128.2.24/23] from ovn-kubernetes multus logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:57:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestapparentmanmyjaeger-1-7dc744766hfj77.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:89629964e44058bafc90393a235b08c7c974f05513aecbfe7229134da732f2b5" already present on machine kubelet logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:57:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestapparentmanmyjaeger-1-7dc744766hfj77.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:57:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestapparentmanmyjaeger-1-7dc744766hfj77.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:57:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestapparentmanmyjaeger-1-7dc744766hfj77.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:a8f53dec42a46c5bf8ac7f2888848c01e70f405d27211e2a269730c888929faf" already present on machine kubelet logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:57:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestapparentmanmyjaeger-1-7dc744766hfj77.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:57:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestapparentmanmyjaeger-1-7dc744766hfj77.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:57:52 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestapparentmanmyjaeger-1-7dc744766hfj77.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:03 +0000 UTC Normal Pod my-jaeger-collector-75c6fdcb99-xv7hw Binding Scheduled Successfully assigned kuttl-test-apparent-man/my-jaeger-collector-75c6fdcb99-xv7hw to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:03 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-75c6fdcb99 SuccessfulCreate Created pod: my-jaeger-collector-75c6fdcb99-xv7hw replicaset-controller logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:03 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-75c6fdcb99 to 1 deployment-controller logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:03 +0000 UTC Normal Pod my-jaeger-query-757889d8c4-2drf9 Binding Scheduled Successfully assigned kuttl-test-apparent-man/my-jaeger-query-757889d8c4-2drf9 to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:03 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-757889d8c4 SuccessfulCreate Created pod: my-jaeger-query-757889d8c4-2drf9 replicaset-controller logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:03 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-757889d8c4 to 1 deployment-controller logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:04 +0000 UTC Normal Pod my-jaeger-collector-75c6fdcb99-xv7hw AddedInterface Add eth0 [10.129.2.31/23] from ovn-kubernetes multus logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:04 +0000 UTC Normal Pod my-jaeger-collector-75c6fdcb99-xv7hw.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:04 +0000 UTC Normal Pod my-jaeger-collector-75c6fdcb99-xv7hw.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:04 +0000 UTC Normal Pod my-jaeger-collector-75c6fdcb99-xv7hw.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:04 +0000 UTC Warning Pod my-jaeger-query-757889d8c4-2drf9 FailedMount MountVolume.SetUp failed for volume "my-jaeger-service-ca" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:04 +0000 UTC Warning Pod my-jaeger-query-757889d8c4-2drf9 FailedMount MountVolume.SetUp failed for volume "certs" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:04 +0000 UTC Warning Pod my-jaeger-query-757889d8c4-2drf9 FailedMount MountVolume.SetUp failed for volume "my-jaeger-ui-configuration-volume" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:04 +0000 UTC Warning Pod my-jaeger-query-757889d8c4-2drf9 FailedMount MountVolume.SetUp failed for volume "kube-api-access-7s8c6" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:05 +0000 UTC Normal Pod my-jaeger-query-757889d8c4-2drf9 AddedInterface Add eth0 [10.131.0.44/23] from ovn-kubernetes multus logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:05 +0000 UTC Normal Pod my-jaeger-query-757889d8c4-2drf9.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:06 +0000 UTC Normal Pod my-jaeger-query-757889d8c4-2drf9.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:06 +0000 UTC Normal Pod my-jaeger-query-757889d8c4-2drf9.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:06 +0000 UTC Normal Pod my-jaeger-query-757889d8c4-2drf9.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:06 +0000 UTC Normal Pod my-jaeger-query-757889d8c4-2drf9.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:06 +0000 UTC Normal Pod my-jaeger-query-757889d8c4-2drf9.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:06 +0000 UTC Normal Pod my-jaeger-query-757889d8c4-2drf9.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:06 +0000 UTC Normal Pod my-jaeger-query-757889d8c4-2drf9.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:06 +0000 UTC Normal Pod my-jaeger-query-757889d8c4-2drf9.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:09 +0000 UTC Normal Pod my-jaeger-collector-75c6fdcb99-xv7hw.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:09 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-75c6fdcb99 SuccessfulDelete Deleted pod: my-jaeger-collector-75c6fdcb99-xv7hw replicaset-controller logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:09 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-75c6fdcb99 to 0 from 1 deployment-controller logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:10 +0000 UTC Normal Pod my-jaeger-collector-79cdc4fc84-qxlrg Binding Scheduled Successfully assigned kuttl-test-apparent-man/my-jaeger-collector-79cdc4fc84-qxlrg to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:10 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-79cdc4fc84 SuccessfulCreate Created pod: my-jaeger-collector-79cdc4fc84-qxlrg replicaset-controller logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:10 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-79cdc4fc84 to 1 deployment-controller logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:11 +0000 UTC Normal Pod my-jaeger-collector-79cdc4fc84-qxlrg AddedInterface Add eth0 [10.129.2.32/23] from ovn-kubernetes multus logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:11 +0000 UTC Normal Pod my-jaeger-collector-79cdc4fc84-qxlrg.spec.containers{jaeger-collector} Pulling Pulling image "test" kubelet logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:11 +0000 UTC Warning Pod my-jaeger-collector-79cdc4fc84-qxlrg.spec.containers{jaeger-collector} Failed Failed to pull image "test": reading manifest latest in docker.io/library/test: requested access to the resource is denied kubelet logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:11 +0000 UTC Warning Pod my-jaeger-collector-79cdc4fc84-qxlrg.spec.containers{jaeger-collector} Failed Error: ErrImagePull kubelet logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:11 +0000 UTC Normal Pod my-jaeger-collector-79cdc4fc84-qxlrg.spec.containers{jaeger-collector} BackOff Back-off pulling image "test" kubelet logger.go:42: 17:58:13 | set-custom-img | 2024-09-30 17:58:11 +0000 UTC Warning Pod my-jaeger-collector-79cdc4fc84-qxlrg.spec.containers{jaeger-collector} Failed Error: ImagePullBackOff kubelet logger.go:42: 17:58:13 | set-custom-img | Deleting namespace: kuttl-test-apparent-man === CONT kuttl/harness/collector-otlp-production-http logger.go:42: 17:58:19 | collector-otlp-production-http | Creating namespace: kuttl-test-kind-anemone logger.go:42: 17:58:20 | collector-otlp-production-http/1-install | starting test step 1-install logger.go:42: 17:58:20 | collector-otlp-production-http/1-install | Jaeger:kuttl-test-kind-anemone/my-jaeger created logger.go:42: 17:58:55 | collector-otlp-production-http/1-install | test step completed 1-install logger.go:42: 17:58:55 | collector-otlp-production-http/2-smoke-test | starting test step 2-smoke-test logger.go:42: 17:58:55 | collector-otlp-production-http/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 17:58:56 | collector-otlp-production-http/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 17:59:03 | collector-otlp-production-http/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 17:59:04 | collector-otlp-production-http/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 17:59:04 | collector-otlp-production-http/2-smoke-test | job.batch/report-span created logger.go:42: 17:59:04 | collector-otlp-production-http/2-smoke-test | job.batch/check-span created logger.go:42: 17:59:16 | collector-otlp-production-http/2-smoke-test | test step completed 2-smoke-test logger.go:42: 17:59:16 | collector-otlp-production-http | collector-otlp-production-http events from ns kuttl-test-kind-anemone: logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestkindanemonemyjaeger-1-77d576f8-chxgk Binding Scheduled Successfully assigned kuttl-test-kind-anemone/elasticsearch-cdm-kuttltestkindanemonemyjaeger-1-77d576f8-chxgk to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestkindanemonemyjaeger-1-77d576f8-chxgk AddedInterface Add eth0 [10.128.2.25/23] from ovn-kubernetes multus logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:26 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestkindanemonemyjaeger-1-77d576f8 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestkindanemonemyjaeger-1-77d576f8-chxgk replicaset-controller logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:26 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestkindanemonemyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestkindanemonemyjaeger-1-77d576f8 to 1 deployment-controller logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:27 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestkindanemonemyjaeger-1-77d576f8-chxgk.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:89629964e44058bafc90393a235b08c7c974f05513aecbfe7229134da732f2b5" already present on machine kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:27 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestkindanemonemyjaeger-1-77d576f8-chxgk.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:27 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestkindanemonemyjaeger-1-77d576f8-chxgk.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:27 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestkindanemonemyjaeger-1-77d576f8-chxgk.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:a8f53dec42a46c5bf8ac7f2888848c01e70f405d27211e2a269730c888929faf" already present on machine kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:27 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestkindanemonemyjaeger-1-77d576f8-chxgk.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:27 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestkindanemonemyjaeger-1-77d576f8-chxgk.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:41 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestkindanemonemyjaeger-1-77d576f8-chxgk.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:53 +0000 UTC Normal Pod my-jaeger-collector-78d7665f5f-vbxxf Binding Scheduled Successfully assigned kuttl-test-kind-anemone/my-jaeger-collector-78d7665f5f-vbxxf to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:53 +0000 UTC Normal Pod my-jaeger-collector-78d7665f5f-vbxxf AddedInterface Add eth0 [10.129.2.33/23] from ovn-kubernetes multus logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:53 +0000 UTC Normal Pod my-jaeger-collector-78d7665f5f-vbxxf.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:53 +0000 UTC Normal Pod my-jaeger-collector-78d7665f5f-vbxxf.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:53 +0000 UTC Normal Pod my-jaeger-collector-78d7665f5f-vbxxf.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:53 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-78d7665f5f SuccessfulCreate Created pod: my-jaeger-collector-78d7665f5f-vbxxf replicaset-controller logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:53 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-78d7665f5f to 1 deployment-controller logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:53 +0000 UTC Normal Pod my-jaeger-query-5bbf669-klzzh Binding Scheduled Successfully assigned kuttl-test-kind-anemone/my-jaeger-query-5bbf669-klzzh to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:53 +0000 UTC Normal Pod my-jaeger-query-5bbf669-klzzh AddedInterface Add eth0 [10.131.0.45/23] from ovn-kubernetes multus logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:53 +0000 UTC Normal Pod my-jaeger-query-5bbf669-klzzh.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:53 +0000 UTC Normal Pod my-jaeger-query-5bbf669-klzzh.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:53 +0000 UTC Normal Pod my-jaeger-query-5bbf669-klzzh.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:53 +0000 UTC Normal Pod my-jaeger-query-5bbf669-klzzh.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:53 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-5bbf669 SuccessfulCreate Created pod: my-jaeger-query-5bbf669-klzzh replicaset-controller logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:53 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-5bbf669 to 1 deployment-controller logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:54 +0000 UTC Normal Pod my-jaeger-query-5bbf669-klzzh.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:54 +0000 UTC Normal Pod my-jaeger-query-5bbf669-klzzh.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:54 +0000 UTC Normal Pod my-jaeger-query-5bbf669-klzzh.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:54 +0000 UTC Normal Pod my-jaeger-query-5bbf669-klzzh.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:54 +0000 UTC Normal Pod my-jaeger-query-5bbf669-klzzh.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:58 +0000 UTC Normal Pod my-jaeger-query-5bbf669-klzzh.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:58 +0000 UTC Normal Pod my-jaeger-query-5bbf669-klzzh.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:58 +0000 UTC Normal Pod my-jaeger-query-5bbf669-klzzh.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:58 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-5bbf669 SuccessfulDelete Deleted pod: my-jaeger-query-5bbf669-klzzh replicaset-controller logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:58 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-5bbf669 to 0 from 1 deployment-controller logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:59 +0000 UTC Normal Pod my-jaeger-query-7d948c4fb7-vrfzc Binding Scheduled Successfully assigned kuttl-test-kind-anemone/my-jaeger-query-7d948c4fb7-vrfzc to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:59 +0000 UTC Normal Pod my-jaeger-query-7d948c4fb7-vrfzc AddedInterface Add eth0 [10.131.0.46/23] from ovn-kubernetes multus logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:59 +0000 UTC Normal Pod my-jaeger-query-7d948c4fb7-vrfzc.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:59 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-7d948c4fb7 SuccessfulCreate Created pod: my-jaeger-query-7d948c4fb7-vrfzc replicaset-controller logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:58:59 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-7d948c4fb7 to 1 deployment-controller logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:59:00 +0000 UTC Normal Pod my-jaeger-query-7d948c4fb7-vrfzc.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:59:00 +0000 UTC Normal Pod my-jaeger-query-7d948c4fb7-vrfzc.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:59:00 +0000 UTC Normal Pod my-jaeger-query-7d948c4fb7-vrfzc.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:59:00 +0000 UTC Normal Pod my-jaeger-query-7d948c4fb7-vrfzc.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:59:00 +0000 UTC Normal Pod my-jaeger-query-7d948c4fb7-vrfzc.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:59:00 +0000 UTC Normal Pod my-jaeger-query-7d948c4fb7-vrfzc.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:59:00 +0000 UTC Normal Pod my-jaeger-query-7d948c4fb7-vrfzc.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:59:00 +0000 UTC Normal Pod my-jaeger-query-7d948c4fb7-vrfzc.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:59:04 +0000 UTC Normal Pod check-span-xbdbm Binding Scheduled Successfully assigned kuttl-test-kind-anemone/check-span-xbdbm to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:59:04 +0000 UTC Normal Pod check-span-xbdbm AddedInterface Add eth0 [10.129.2.35/23] from ovn-kubernetes multus logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:59:04 +0000 UTC Normal Pod check-span-xbdbm.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:59:04 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-xbdbm job-controller logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:59:04 +0000 UTC Normal Pod report-span-glnf8 Binding Scheduled Successfully assigned kuttl-test-kind-anemone/report-span-glnf8 to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:59:04 +0000 UTC Normal Pod report-span-glnf8 AddedInterface Add eth0 [10.129.2.34/23] from ovn-kubernetes multus logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:59:04 +0000 UTC Normal Pod report-span-glnf8.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:59:04 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-glnf8 job-controller logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:59:05 +0000 UTC Normal Pod check-span-xbdbm.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 216ms (216ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:59:05 +0000 UTC Normal Pod check-span-xbdbm.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:59:05 +0000 UTC Normal Pod check-span-xbdbm.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:59:05 +0000 UTC Normal Pod report-span-glnf8.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 235ms (235ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:59:05 +0000 UTC Normal Pod report-span-glnf8.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:59:05 +0000 UTC Normal Pod report-span-glnf8.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:59:08 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:59:08 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:59:08 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 17:59:16 | collector-otlp-production-http | 2024-09-30 17:59:15 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 17:59:16 | collector-otlp-production-http | Deleting namespace: kuttl-test-kind-anemone === CONT kuttl/harness/collector-otlp-production-grpc logger.go:42: 17:59:29 | collector-otlp-production-grpc | Creating namespace: kuttl-test-touched-grubworm logger.go:42: 17:59:29 | collector-otlp-production-grpc/1-install | starting test step 1-install logger.go:42: 17:59:29 | collector-otlp-production-grpc/1-install | Jaeger:kuttl-test-touched-grubworm/my-jaeger created logger.go:42: 18:00:04 | collector-otlp-production-grpc/1-install | test step completed 1-install logger.go:42: 18:00:04 | collector-otlp-production-grpc/2-smoke-test | starting test step 2-smoke-test logger.go:42: 18:00:04 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 18:00:05 | collector-otlp-production-grpc/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 18:00:13 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 18:00:14 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 18:00:14 | collector-otlp-production-grpc/2-smoke-test | job.batch/report-span created logger.go:42: 18:00:14 | collector-otlp-production-grpc/2-smoke-test | job.batch/check-span created logger.go:42: 18:00:34 | collector-otlp-production-grpc/2-smoke-test | test step completed 2-smoke-test logger.go:42: 18:00:34 | collector-otlp-production-grpc | collector-otlp-production-grpc events from ns kuttl-test-touched-grubworm: logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 17:59:34 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltesttouchedgrubwormmyjaeger-1-8575888566 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltesttouchedgrubwormmyjaeger-1-85758d5wbs replicaset-controller logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 17:59:34 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttouchedgrubwormmyjaeger-1-85758d5wbs Binding Scheduled Successfully assigned kuttl-test-touched-grubworm/elasticsearch-cdm-kuttltesttouchedgrubwormmyjaeger-1-85758d5wbs to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 17:59:34 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltesttouchedgrubwormmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltesttouchedgrubwormmyjaeger-1-8575888566 to 1 deployment-controller logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 17:59:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttouchedgrubwormmyjaeger-1-85758d5wbs AddedInterface Add eth0 [10.128.2.26/23] from ovn-kubernetes multus logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 17:59:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttouchedgrubwormmyjaeger-1-85758d5wbs.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:89629964e44058bafc90393a235b08c7c974f05513aecbfe7229134da732f2b5" already present on machine kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 17:59:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttouchedgrubwormmyjaeger-1-85758d5wbs.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 17:59:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttouchedgrubwormmyjaeger-1-85758d5wbs.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 17:59:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttouchedgrubwormmyjaeger-1-85758d5wbs.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:a8f53dec42a46c5bf8ac7f2888848c01e70f405d27211e2a269730c888929faf" already present on machine kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 17:59:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttouchedgrubwormmyjaeger-1-85758d5wbs.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 17:59:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttouchedgrubwormmyjaeger-1-85758d5wbs.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 17:59:45 +0000 UTC Warning Pod elasticsearch-cdm-kuttltesttouchedgrubwormmyjaeger-1-85758d5wbs.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 17:59:50 +0000 UTC Warning Pod elasticsearch-cdm-kuttltesttouchedgrubwormmyjaeger-1-85758d5wbs.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:01 +0000 UTC Normal Pod my-jaeger-collector-7d94d6747-b6msh Binding Scheduled Successfully assigned kuttl-test-touched-grubworm/my-jaeger-collector-7d94d6747-b6msh to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:01 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7d94d6747 SuccessfulCreate Created pod: my-jaeger-collector-7d94d6747-b6msh replicaset-controller logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:01 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-7d94d6747 to 1 deployment-controller logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:01 +0000 UTC Normal Pod my-jaeger-query-8d994d5f8-z2m85 Binding Scheduled Successfully assigned kuttl-test-touched-grubworm/my-jaeger-query-8d994d5f8-z2m85 to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:01 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-8d994d5f8 SuccessfulCreate Created pod: my-jaeger-query-8d994d5f8-z2m85 replicaset-controller logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:01 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-8d994d5f8 to 1 deployment-controller logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:02 +0000 UTC Normal Pod my-jaeger-collector-7d94d6747-b6msh AddedInterface Add eth0 [10.129.2.36/23] from ovn-kubernetes multus logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:02 +0000 UTC Normal Pod my-jaeger-collector-7d94d6747-b6msh.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:02 +0000 UTC Normal Pod my-jaeger-collector-7d94d6747-b6msh.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:02 +0000 UTC Normal Pod my-jaeger-collector-7d94d6747-b6msh.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:02 +0000 UTC Normal Pod my-jaeger-query-8d994d5f8-z2m85 AddedInterface Add eth0 [10.131.0.48/23] from ovn-kubernetes multus logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:02 +0000 UTC Normal Pod my-jaeger-query-8d994d5f8-z2m85.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:02 +0000 UTC Normal Pod my-jaeger-query-8d994d5f8-z2m85.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:02 +0000 UTC Normal Pod my-jaeger-query-8d994d5f8-z2m85.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:02 +0000 UTC Normal Pod my-jaeger-query-8d994d5f8-z2m85.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:02 +0000 UTC Normal Pod my-jaeger-query-8d994d5f8-z2m85.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:02 +0000 UTC Normal Pod my-jaeger-query-8d994d5f8-z2m85.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:02 +0000 UTC Normal Pod my-jaeger-query-8d994d5f8-z2m85.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:02 +0000 UTC Normal Pod my-jaeger-query-8d994d5f8-z2m85.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:02 +0000 UTC Normal Pod my-jaeger-query-8d994d5f8-z2m85.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:10 +0000 UTC Normal Pod my-jaeger-query-8d994d5f8-z2m85.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:10 +0000 UTC Normal Pod my-jaeger-query-8d994d5f8-z2m85.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:10 +0000 UTC Normal Pod my-jaeger-query-8d994d5f8-z2m85.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:10 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-8d994d5f8 SuccessfulDelete Deleted pod: my-jaeger-query-8d994d5f8-z2m85 replicaset-controller logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:10 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-8d994d5f8 to 0 from 1 deployment-controller logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:11 +0000 UTC Normal Pod my-jaeger-query-57875c4cd6-p9ffm Binding Scheduled Successfully assigned kuttl-test-touched-grubworm/my-jaeger-query-57875c4cd6-p9ffm to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:11 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-57875c4cd6 SuccessfulCreate Created pod: my-jaeger-query-57875c4cd6-p9ffm replicaset-controller logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:11 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-57875c4cd6 to 1 deployment-controller logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:12 +0000 UTC Normal Pod my-jaeger-query-57875c4cd6-p9ffm AddedInterface Add eth0 [10.131.0.49/23] from ovn-kubernetes multus logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:12 +0000 UTC Normal Pod my-jaeger-query-57875c4cd6-p9ffm.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:12 +0000 UTC Normal Pod my-jaeger-query-57875c4cd6-p9ffm.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:12 +0000 UTC Normal Pod my-jaeger-query-57875c4cd6-p9ffm.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:12 +0000 UTC Normal Pod my-jaeger-query-57875c4cd6-p9ffm.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:12 +0000 UTC Normal Pod my-jaeger-query-57875c4cd6-p9ffm.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:12 +0000 UTC Normal Pod my-jaeger-query-57875c4cd6-p9ffm.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:12 +0000 UTC Normal Pod my-jaeger-query-57875c4cd6-p9ffm.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:12 +0000 UTC Normal Pod my-jaeger-query-57875c4cd6-p9ffm.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:12 +0000 UTC Normal Pod my-jaeger-query-57875c4cd6-p9ffm.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:14 +0000 UTC Normal Pod check-span-xm6mp Binding Scheduled Successfully assigned kuttl-test-touched-grubworm/check-span-xm6mp to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:14 +0000 UTC Normal Pod check-span-xm6mp AddedInterface Add eth0 [10.129.2.38/23] from ovn-kubernetes multus logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:14 +0000 UTC Normal Pod check-span-xm6mp.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:14 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-xm6mp job-controller logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:14 +0000 UTC Normal Pod report-span-ttrl7 Binding Scheduled Successfully assigned kuttl-test-touched-grubworm/report-span-ttrl7 to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:14 +0000 UTC Normal Pod report-span-ttrl7 AddedInterface Add eth0 [10.129.2.37/23] from ovn-kubernetes multus logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:14 +0000 UTC Normal Pod report-span-ttrl7.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:14 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-ttrl7 job-controller logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:15 +0000 UTC Normal Pod check-span-xm6mp.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 283ms (283ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:15 +0000 UTC Normal Pod check-span-xm6mp.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:15 +0000 UTC Normal Pod check-span-xm6mp.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:15 +0000 UTC Normal Pod report-span-ttrl7.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 273ms (273ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:15 +0000 UTC Normal Pod report-span-ttrl7.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:15 +0000 UTC Normal Pod report-span-ttrl7.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:16 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:16 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:16 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 18:00:34 | collector-otlp-production-grpc | 2024-09-30 18:00:34 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 18:00:34 | collector-otlp-production-grpc | Deleting namespace: kuttl-test-touched-grubworm === CONT kuttl/harness/collector-autoscale logger.go:42: 18:00:41 | collector-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 18:00:41 | collector-autoscale | Creating namespace: kuttl-test-patient-urchin logger.go:42: 18:00:42 | collector-autoscale/1-install | starting test step 1-install logger.go:42: 18:00:42 | collector-autoscale/1-install | Jaeger:kuttl-test-patient-urchin/simple-prod created logger.go:42: 18:01:19 | collector-autoscale/1-install | test step completed 1-install logger.go:42: 18:01:19 | collector-autoscale/2- | starting test step 2- logger.go:42: 18:01:19 | collector-autoscale/2- | test step completed 2- logger.go:42: 18:01:19 | collector-autoscale | collector-autoscale events from ns kuttl-test-patient-urchin: logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:00:49 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestpatienturchinsimpleprod-1-6654d8c9c6 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestpatienturchinsimpleprod-1-6654drpqsg replicaset-controller logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:00:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpatienturchinsimpleprod-1-6654drpqsg Binding Scheduled Successfully assigned kuttl-test-patient-urchin/elasticsearch-cdm-kuttltestpatienturchinsimpleprod-1-6654drpqsg to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:00:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpatienturchinsimpleprod-1-6654drpqsg AddedInterface Add eth0 [10.128.2.27/23] from ovn-kubernetes multus logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:00:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpatienturchinsimpleprod-1-6654drpqsg.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:89629964e44058bafc90393a235b08c7c974f05513aecbfe7229134da732f2b5" already present on machine kubelet logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:00:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpatienturchinsimpleprod-1-6654drpqsg.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:00:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpatienturchinsimpleprod-1-6654drpqsg.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:00:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpatienturchinsimpleprod-1-6654drpqsg.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:a8f53dec42a46c5bf8ac7f2888848c01e70f405d27211e2a269730c888929faf" already present on machine kubelet logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:00:49 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestpatienturchinsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestpatienturchinsimpleprod-1-6654d8c9c6 to 1 deployment-controller logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:00:50 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpatienturchinsimpleprod-1-6654drpqsg.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:00:50 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpatienturchinsimpleprod-1-6654drpqsg.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:00:59 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestpatienturchinsimpleprod-1-6654drpqsg.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:01:04 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestpatienturchinsimpleprod-1-6654drpqsg.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:01:15 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-6d7b955486 to 1 deployment-controller logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:01:16 +0000 UTC Normal Pod simple-prod-collector-6d7b955486-nvcgw Binding Scheduled Successfully assigned kuttl-test-patient-urchin/simple-prod-collector-6d7b955486-nvcgw to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:01:16 +0000 UTC Normal Pod simple-prod-collector-6d7b955486-nvcgw AddedInterface Add eth0 [10.129.2.39/23] from ovn-kubernetes multus logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:01:16 +0000 UTC Normal Pod simple-prod-collector-6d7b955486-nvcgw.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:01:16 +0000 UTC Normal Pod simple-prod-collector-6d7b955486-nvcgw.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:01:16 +0000 UTC Normal Pod simple-prod-collector-6d7b955486-nvcgw.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:01:16 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-6d7b955486 SuccessfulCreate Created pod: simple-prod-collector-6d7b955486-nvcgw replicaset-controller logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:01:16 +0000 UTC Normal Pod simple-prod-query-5c4756f7d-h5lz7 Binding Scheduled Successfully assigned kuttl-test-patient-urchin/simple-prod-query-5c4756f7d-h5lz7 to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:01:16 +0000 UTC Warning Pod simple-prod-query-5c4756f7d-h5lz7 FailedMount MountVolume.SetUp failed for volume "simple-prod-ui-oauth-proxy-tls" : secret "simple-prod-ui-oauth-proxy-tls" not found kubelet logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:01:16 +0000 UTC Normal ReplicaSet.apps simple-prod-query-5c4756f7d SuccessfulCreate Created pod: simple-prod-query-5c4756f7d-h5lz7 replicaset-controller logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:01:16 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-5c4756f7d to 1 deployment-controller logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:01:17 +0000 UTC Normal Pod simple-prod-query-5c4756f7d-h5lz7 AddedInterface Add eth0 [10.131.0.50/23] from ovn-kubernetes multus logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:01:17 +0000 UTC Normal Pod simple-prod-query-5c4756f7d-h5lz7.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:01:17 +0000 UTC Normal Pod simple-prod-query-5c4756f7d-h5lz7.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:01:17 +0000 UTC Normal Pod simple-prod-query-5c4756f7d-h5lz7.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:01:17 +0000 UTC Normal Pod simple-prod-query-5c4756f7d-h5lz7.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:01:17 +0000 UTC Normal Pod simple-prod-query-5c4756f7d-h5lz7.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:01:17 +0000 UTC Normal Pod simple-prod-query-5c4756f7d-h5lz7.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:01:17 +0000 UTC Normal Pod simple-prod-query-5c4756f7d-h5lz7.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:01:17 +0000 UTC Normal Pod simple-prod-query-5c4756f7d-h5lz7.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 18:01:19 | collector-autoscale | 2024-09-30 18:01:17 +0000 UTC Normal Pod simple-prod-query-5c4756f7d-h5lz7.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 18:01:19 | collector-autoscale | Deleting namespace: kuttl-test-patient-urchin === CONT kuttl/harness/collector-otlp-allinone-grpc logger.go:42: 18:01:25 | collector-otlp-allinone-grpc | Creating namespace: kuttl-test-definite-redfish logger.go:42: 18:01:25 | collector-otlp-allinone-grpc/0-install | starting test step 0-install logger.go:42: 18:01:25 | collector-otlp-allinone-grpc/0-install | Jaeger:kuttl-test-definite-redfish/my-jaeger created logger.go:42: 18:01:31 | collector-otlp-allinone-grpc/0-install | test step completed 0-install logger.go:42: 18:01:31 | collector-otlp-allinone-grpc/1-smoke-test | starting test step 1-smoke-test logger.go:42: 18:01:31 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 18:01:33 | collector-otlp-allinone-grpc/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 18:01:39 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 18:01:40 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 18:01:40 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/report-span created logger.go:42: 18:01:40 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/check-span created logger.go:42: 18:02:00 | collector-otlp-allinone-grpc/1-smoke-test | test step completed 1-smoke-test logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | collector-otlp-allinone-grpc events from ns kuttl-test-definite-redfish: logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:29 +0000 UTC Normal Pod my-jaeger-794b94d4d7-xk5qt Binding Scheduled Successfully assigned kuttl-test-definite-redfish/my-jaeger-794b94d4d7-xk5qt to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:29 +0000 UTC Normal ReplicaSet.apps my-jaeger-794b94d4d7 SuccessfulCreate Created pod: my-jaeger-794b94d4d7-xk5qt replicaset-controller logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:29 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-794b94d4d7 to 1 deployment-controller logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:30 +0000 UTC Normal Pod my-jaeger-794b94d4d7-xk5qt AddedInterface Add eth0 [10.131.0.51/23] from ovn-kubernetes multus logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:30 +0000 UTC Normal Pod my-jaeger-794b94d4d7-xk5qt.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:30 +0000 UTC Normal Pod my-jaeger-794b94d4d7-xk5qt.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:30 +0000 UTC Normal Pod my-jaeger-794b94d4d7-xk5qt.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:30 +0000 UTC Normal Pod my-jaeger-794b94d4d7-xk5qt.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:30 +0000 UTC Normal Pod my-jaeger-794b94d4d7-xk5qt.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:30 +0000 UTC Normal Pod my-jaeger-794b94d4d7-xk5qt.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:34 +0000 UTC Normal Pod my-jaeger-794b94d4d7-xk5qt.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:34 +0000 UTC Normal Pod my-jaeger-794b94d4d7-xk5qt.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:34 +0000 UTC Normal ReplicaSet.apps my-jaeger-794b94d4d7 SuccessfulDelete Deleted pod: my-jaeger-794b94d4d7-xk5qt replicaset-controller logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:34 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-794b94d4d7 to 0 from 1 deployment-controller logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:35 +0000 UTC Normal Pod my-jaeger-57467f45c6-kmhtq Binding Scheduled Successfully assigned kuttl-test-definite-redfish/my-jaeger-57467f45c6-kmhtq to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:35 +0000 UTC Normal ReplicaSet.apps my-jaeger-57467f45c6 SuccessfulCreate Created pod: my-jaeger-57467f45c6-kmhtq replicaset-controller logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:35 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-57467f45c6 to 1 deployment-controller logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:36 +0000 UTC Normal Pod my-jaeger-57467f45c6-kmhtq AddedInterface Add eth0 [10.131.0.52/23] from ovn-kubernetes multus logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:36 +0000 UTC Normal Pod my-jaeger-57467f45c6-kmhtq.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:36 +0000 UTC Normal Pod my-jaeger-57467f45c6-kmhtq.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:36 +0000 UTC Normal Pod my-jaeger-57467f45c6-kmhtq.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:36 +0000 UTC Normal Pod my-jaeger-57467f45c6-kmhtq.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:36 +0000 UTC Normal Pod my-jaeger-57467f45c6-kmhtq.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:36 +0000 UTC Normal Pod my-jaeger-57467f45c6-kmhtq.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:40 +0000 UTC Normal Pod check-span-z76ll Binding Scheduled Successfully assigned kuttl-test-definite-redfish/check-span-z76ll to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:40 +0000 UTC Normal Pod check-span-z76ll AddedInterface Add eth0 [10.129.2.40/23] from ovn-kubernetes multus logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:40 +0000 UTC Normal Pod check-span-z76ll.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:40 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-z76ll job-controller logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:40 +0000 UTC Normal Pod report-span-jk95r Binding Scheduled Successfully assigned kuttl-test-definite-redfish/report-span-jk95r to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:40 +0000 UTC Normal Pod report-span-jk95r AddedInterface Add eth0 [10.128.2.28/23] from ovn-kubernetes multus logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:40 +0000 UTC Normal Pod report-span-jk95r.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:40 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-jk95r job-controller logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:41 +0000 UTC Normal Pod check-span-z76ll.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 216ms (216ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:41 +0000 UTC Normal Pod check-span-z76ll.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:41 +0000 UTC Normal Pod check-span-z76ll.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:42 +0000 UTC Normal Pod report-span-jk95r.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.166s (1.166s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:42 +0000 UTC Normal Pod report-span-jk95r.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:01:42 +0000 UTC Normal Pod report-span-jk95r.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | 2024-09-30 18:02:00 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 18:02:00 | collector-otlp-allinone-grpc | Deleting namespace: kuttl-test-definite-redfish === CONT kuttl/harness/cassandra-spark logger.go:42: 18:02:12 | cassandra-spark | Ignoring 01-assert.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 18:02:12 | cassandra-spark | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 18:02:12 | cassandra-spark | Creating namespace: kuttl-test-notable-unicorn logger.go:42: 18:02:12 | cassandra-spark | cassandra-spark events from ns kuttl-test-notable-unicorn: logger.go:42: 18:02:12 | cassandra-spark | Deleting namespace: kuttl-test-notable-unicorn === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (335.10s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.21s) --- PASS: kuttl/harness/collector-otlp-allinone-http (39.86s) --- PASS: kuttl/harness/set-custom-img (50.14s) --- PASS: kuttl/harness/collector-otlp-production-http (69.29s) --- PASS: kuttl/harness/collector-otlp-production-grpc (72.70s) --- PASS: kuttl/harness/collector-autoscale (43.64s) --- PASS: kuttl/harness/collector-otlp-allinone-grpc (47.15s) --- PASS: kuttl/harness/cassandra-spark (5.92s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name miscellaneous --report --output /logs/artifacts/miscellaneous.xml ./artifacts/kuttl-report.xml time="2024-09-30T18:02:18Z" level=debug msg="Setting a new name for the test suites" time="2024-09-30T18:02:18Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-09-30T18:02:18Z" level=debug msg="normalizing test case names" time="2024-09-30T18:02:18Z" level=debug msg="miscellaneous/artifacts -> miscellaneous_artifacts" time="2024-09-30T18:02:18Z" level=debug msg="miscellaneous/collector-otlp-allinone-http -> miscellaneous_collector_otlp_allinone_http" time="2024-09-30T18:02:18Z" level=debug msg="miscellaneous/set-custom-img -> miscellaneous_set_custom_img" time="2024-09-30T18:02:18Z" level=debug msg="miscellaneous/collector-otlp-production-http -> miscellaneous_collector_otlp_production_http" time="2024-09-30T18:02:18Z" level=debug msg="miscellaneous/collector-otlp-production-grpc -> miscellaneous_collector_otlp_production_grpc" time="2024-09-30T18:02:18Z" level=debug msg="miscellaneous/collector-autoscale -> miscellaneous_collector_autoscale" time="2024-09-30T18:02:18Z" level=debug msg="miscellaneous/collector-otlp-allinone-grpc -> miscellaneous_collector_otlp_allinone_grpc" time="2024-09-30T18:02:18Z" level=debug msg="miscellaneous/cassandra-spark -> miscellaneous_cassandra_spark" +----------------------------------------------+--------+ | NAME | RESULT | +----------------------------------------------+--------+ | miscellaneous_artifacts | passed | | miscellaneous_collector_otlp_allinone_http | passed | | miscellaneous_set_custom_img | passed | | miscellaneous_collector_otlp_production_http | passed | | miscellaneous_collector_otlp_production_grpc | passed | | miscellaneous_collector_autoscale | passed | | miscellaneous_collector_otlp_allinone_grpc | passed | | miscellaneous_cassandra_spark | passed | +----------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh elasticsearch false true + '[' 3 -ne 3 ']' + test_suite_name=elasticsearch + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/elasticsearch.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-elasticsearch make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true \ KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ ./tests/e2e/elasticsearch/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.17.0-0.nightly-2024-09-29-173421 True False 25m Cluster version is 4.17.0-0.nightly-2024-09-29-173421' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.17.0-0.nightly-2024-09-29-173421 True False 25m Cluster version is 4.17.0-0.nightly-2024-09-29-173421' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/elasticsearch/render.sh ++ export SUITE_DIR=./tests/e2e/elasticsearch ++ SUITE_DIR=./tests/e2e/elasticsearch ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/elasticsearch ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + start_test es-from-aio-to-production + '[' 1 -ne 1 ']' + test_name=es-from-aio-to-production + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-from-aio-to-production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-from-aio-to-production\e[0m' Rendering files for test es-from-aio-to-production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-from-aio-to-production + cd es-from-aio-to-production + jaeger_name=my-jaeger + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 03 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=03 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./03-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch.redundancyPolicy="ZeroRedundancy"' ./03-install.yaml + render_smoke_test my-jaeger true 04 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=04 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./04-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./04-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test es-increasing-replicas + '[' 1 -ne 1 ']' + test_name=es-increasing-replicas + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-increasing-replicas' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-increasing-replicas\e[0m' Rendering files for test es-increasing-replicas + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-from-aio-to-production + '[' es-from-aio-to-production '!=' _build ']' + cd .. + mkdir -p es-increasing-replicas + cd es-increasing-replicas + jaeger_name=simple-prod + '[' true = true ']' + jaeger_deployment_mode=production_autoprovisioned + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.replicas=2 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.query.replicas=2 ./02-install.yaml + cp ./01-assert.yaml ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.replicas=2 ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .status.readyReplicas=2 ./02-assert.yaml + render_smoke_test simple-prod true 03 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=03 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./03-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + cp ./02-install.yaml ./04-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.elasticsearch.nodeCount=2 ./04-install.yaml + /tmp/jaeger-tests/bin/gomplate -f ./openshift-check-es-nodes.yaml.template -o ./05-check-es-nodes.yaml + '[' true = true ']' + skip_test es-index-cleaner-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-increasing-replicas + '[' es-increasing-replicas '!=' _build ']' + cd .. + rm -rf es-index-cleaner-upstream + warning 'es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_index_cleaner -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-index-cleaner-autoprov + '[' 1 -ne 1 ']' + test_name=es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-index-cleaner-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-index-cleaner-autoprov\e[0m' Rendering files for test es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-index-cleaner-autoprov + cd es-index-cleaner-autoprov + jaeger_name=test-es-index-cleaner-with-prefix + cronjob_name=test-es-index-cleaner-with-prefix-es-index-cleaner + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + cp ../../es-index-cleaner-upstream/04-assert.yaml ../../es-index-cleaner-upstream/README.md . + render_install_jaeger test-es-index-cleaner-with-prefix production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options.es.index-prefix=""' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.enabled=false ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.numberOfDays=0 ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.esIndexCleaner.schedule="*/1 * * * *"' ./01-install.yaml + render_report_spans test-es-index-cleaner-with-prefix true 5 00 true 02 + '[' 6 -ne 6 ']' + jaeger=test-es-index-cleaner-with-prefix + is_secured=true + number_of_spans=5 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + export JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=5 + DAYS=5 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + sed 's~enabled: false~enabled: true~gi' ./01-install.yaml + CRONJOB_NAME=test-es-index-cleaner-with-prefix-es-index-cleaner + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./04-wait-es-index-cleaner.yaml + /tmp/jaeger-tests/bin/gomplate -f ./01-install.yaml -o ./05-install.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 00 06 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=00 + test_step=06 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=test-es-index-cleaner-with-prefix-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=test-es-index-cleaner-with-prefix-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./06-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./06-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.13"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.13"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.13"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.13 ++ version_ge 5.8.13 5.4 +++ echo 5.8.13 5.4 +++ tr ' ' '\n' +++ sort -rV +++ head -n 1 ++ test 5.8.13 == 5.8.13 + '[' -n '' ']' + skip_test es-index-cleaner-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-index-cleaner-autoprov + '[' es-index-cleaner-autoprov '!=' _build ']' + cd .. + rm -rf es-index-cleaner-managed + warning 'es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + start_test es-multiinstance + '[' 1 -ne 1 ']' + test_name=es-multiinstance + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-multiinstance' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-multiinstance\e[0m' Rendering files for test es-multiinstance + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-multiinstance + cd es-multiinstance + jaeger_name=instance-1 + render_install_jaeger instance-1 production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=instance-1 + JAEGER_NAME=instance-1 + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f ./03-create-second-instance.yaml.template -o 03-create-second-instance.yaml + '[' true = true ']' + skip_test es-rollover-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-rollover-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-multiinstance + '[' es-multiinstance '!=' _build ']' + cd .. + rm -rf es-rollover-upstream + warning 'es-rollover-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_rollover -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-rollover-autoprov + '[' 1 -ne 1 ']' + test_name=es-rollover-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-rollover-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-rollover-autoprov\e[0m' Rendering files for test es-rollover-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-rollover-autoprov + cd es-rollover-autoprov + cp ../../es-rollover-upstream/05-assert.yaml ../../es-rollover-upstream/05-install.yaml ../../es-rollover-upstream/README.md . + jaeger_name=my-jaeger + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_report_spans my-jaeger true 2 00 true 02 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 00 03 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=00 + test_step=03 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./03-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./03-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 01 04 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=01 + test_step=04 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=01 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./04-check-indices.yaml + JOB_NUMBER=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./04-assert.yaml + render_report_spans my-jaeger true 2 02 true 06 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=02 + ensure_reported_spans=true + test_step=06 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=02 + JOB_NUMBER=02 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./06-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./06-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 02 07 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=02 + test_step=07 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=02 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./07-check-indices.yaml + JOB_NUMBER=02 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./07-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' 03 08 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + job_number=03 + test_step=08 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=03 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./08-check-indices.yaml + JOB_NUMBER=03 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./08-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' 04 09 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + job_number=04 + test_step=09 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=04 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./09-check-indices.yaml + JOB_NUMBER=04 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./09-assert.yaml + render_report_spans my-jaeger true 2 03 true 10 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=03 + ensure_reported_spans=true + test_step=10 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=03 + JOB_NUMBER=03 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./10-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./10-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + CRONJOB_NAME=my-jaeger-es-rollover + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./11-wait-rollover.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-000002'\'',' 05 11 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-000002'\'',' + job_number=05 + test_step=11 + escape_command ''\''--name'\'', '\''jaeger-span-000002'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-000002'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-000002'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-000002'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=05 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./11-check-indices.yaml + JOB_NUMBER=05 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./11-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' 06 12 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + job_number=06 + test_step=12 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=06 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./12-check-indices.yaml + JOB_NUMBER=06 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./12-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.13"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.13"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.13"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.13 ++ version_ge 5.8.13 5.4 +++ echo 5.8.13 5.4 +++ tr ' ' '\n' +++ sort -rV +++ head -n 1 ++ test 5.8.13 == 5.8.13 + '[' -n '' ']' + skip_test es-rollover-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-rollover-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-rollover-autoprov + '[' es-rollover-autoprov '!=' _build ']' + cd .. + rm -rf es-rollover-managed + warning 'es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + skip_test es-spark-dependencies 'This test is not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=es-spark-dependencies + message='This test is not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + rm -rf es-spark-dependencies + warning 'es-spark-dependencies: This test is not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-spark-dependencies: This test is not supported in OpenShift\e[0m' WAR: es-spark-dependencies: This test is not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running elasticsearch E2E tests' Running elasticsearch E2E tests + cd tests/e2e/elasticsearch/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2856073267 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 7 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/es-from-aio-to-production === PAUSE kuttl/harness/es-from-aio-to-production === RUN kuttl/harness/es-increasing-replicas === PAUSE kuttl/harness/es-increasing-replicas === RUN kuttl/harness/es-index-cleaner-autoprov === PAUSE kuttl/harness/es-index-cleaner-autoprov === RUN kuttl/harness/es-multiinstance === PAUSE kuttl/harness/es-multiinstance === RUN kuttl/harness/es-rollover-autoprov === PAUSE kuttl/harness/es-rollover-autoprov === RUN kuttl/harness/es-simple-prod === PAUSE kuttl/harness/es-simple-prod === CONT kuttl/harness/artifacts logger.go:42: 18:02:45 | artifacts | Creating namespace: kuttl-test-prepared-goblin logger.go:42: 18:02:45 | artifacts | artifacts events from ns kuttl-test-prepared-goblin: logger.go:42: 18:02:45 | artifacts | Deleting namespace: kuttl-test-prepared-goblin === CONT kuttl/harness/es-multiinstance logger.go:42: 18:02:51 | es-multiinstance | Ignoring 03-create-second-instance.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 18:02:51 | es-multiinstance | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 18:02:51 | es-multiinstance | Creating namespace: kuttl-test-improved-koi logger.go:42: 18:02:51 | es-multiinstance/0-clear-namespace | starting test step 0-clear-namespace logger.go:42: 18:02:51 | es-multiinstance/0-clear-namespace | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --ignore-not-found=true] logger.go:42: 18:02:51 | es-multiinstance/0-clear-namespace | test step completed 0-clear-namespace logger.go:42: 18:02:51 | es-multiinstance/1-install | starting test step 1-install logger.go:42: 18:02:51 | es-multiinstance/1-install | Jaeger:kuttl-test-improved-koi/instance-1 created logger.go:42: 18:03:30 | es-multiinstance/1-install | test step completed 1-install logger.go:42: 18:03:30 | es-multiinstance/2-create-namespace | starting test step 2-create-namespace logger.go:42: 18:03:30 | es-multiinstance/2-create-namespace | running command: [sh -c kubectl create namespace jaeger-e2e-multiinstance-test] logger.go:42: 18:03:30 | es-multiinstance/2-create-namespace | namespace/jaeger-e2e-multiinstance-test created logger.go:42: 18:03:30 | es-multiinstance/2-create-namespace | test step completed 2-create-namespace logger.go:42: 18:03:30 | es-multiinstance/3-create-second-instance | starting test step 3-create-second-instance logger.go:42: 18:03:30 | es-multiinstance/3-create-second-instance | running command: [sh -c kubectl apply -f ./01-install.yaml -n jaeger-e2e-multiinstance-test] logger.go:42: 18:03:30 | es-multiinstance/3-create-second-instance | jaeger.jaegertracing.io/instance-1 created logger.go:42: 18:03:30 | es-multiinstance/3-create-second-instance | running command: [sh -c /tmp/jaeger-tests/bin/kubectl-kuttl assert ./01-assert.yaml -n jaeger-e2e-multiinstance-test --timeout 1000] logger.go:42: 18:04:21 | es-multiinstance/3-create-second-instance | assert is valid logger.go:42: 18:04:21 | es-multiinstance/3-create-second-instance | test step completed 3-create-second-instance logger.go:42: 18:04:21 | es-multiinstance/4-check-secrets | starting test step 4-check-secrets logger.go:42: 18:04:21 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n $NAMESPACE > secret1] logger.go:42: 18:04:21 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n jaeger-e2e-multiinstance-test > secret2] logger.go:42: 18:04:21 | es-multiinstance/4-check-secrets | running command: [sh -c cmp --silent secret1 secret2 || exit 0] logger.go:42: 18:04:21 | es-multiinstance/4-check-secrets | test step completed 4-check-secrets logger.go:42: 18:04:21 | es-multiinstance/5-delete | starting test step 5-delete logger.go:42: 18:04:21 | es-multiinstance/5-delete | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --wait=false] logger.go:42: 18:04:21 | es-multiinstance/5-delete | namespace "jaeger-e2e-multiinstance-test" deleted logger.go:42: 18:04:21 | es-multiinstance/5-delete | test step completed 5-delete logger.go:42: 18:04:21 | es-multiinstance | es-multiinstance events from ns kuttl-test-improved-koi: logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:00 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestimprovedkoiinstance1-1-7c68955d8 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestimprovedkoiinstance1-1-7c68955dcg8h5 replicaset-controller logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:00 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestimprovedkoiinstance1-1-7c68955dcg8h5 Binding Scheduled Successfully assigned kuttl-test-improved-koi/elasticsearch-cdm-kuttltestimprovedkoiinstance1-1-7c68955dcg8h5 to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:00 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestimprovedkoiinstance1-1-7c68955dcg8h5 FailedMount MountVolume.SetUp failed for volume "elasticsearch-metrics" : secret "elasticsearch-metrics" not found kubelet logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:00 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestimprovedkoiinstance1-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestimprovedkoiinstance1-1-7c68955d8 to 1 deployment-controller logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:01 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestimprovedkoiinstance1-1-7c68955dcg8h5 AddedInterface Add eth0 [10.128.2.29/23] from ovn-kubernetes multus logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:01 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestimprovedkoiinstance1-1-7c68955dcg8h5.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:89629964e44058bafc90393a235b08c7c974f05513aecbfe7229134da732f2b5" already present on machine kubelet logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:01 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestimprovedkoiinstance1-1-7c68955dcg8h5.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:01 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestimprovedkoiinstance1-1-7c68955dcg8h5.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:01 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestimprovedkoiinstance1-1-7c68955dcg8h5.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:a8f53dec42a46c5bf8ac7f2888848c01e70f405d27211e2a269730c888929faf" already present on machine kubelet logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:01 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestimprovedkoiinstance1-1-7c68955dcg8h5.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:01 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestimprovedkoiinstance1-1-7c68955dcg8h5.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:11 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestimprovedkoiinstance1-1-7c68955dcg8h5.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:16 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestimprovedkoiinstance1-1-7c68955dcg8h5.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:27 +0000 UTC Normal Pod instance-1-collector-545ff6d759-mpgt5 Binding Scheduled Successfully assigned kuttl-test-improved-koi/instance-1-collector-545ff6d759-mpgt5 to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:27 +0000 UTC Normal ReplicaSet.apps instance-1-collector-545ff6d759 SuccessfulCreate Created pod: instance-1-collector-545ff6d759-mpgt5 replicaset-controller logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:27 +0000 UTC Normal Deployment.apps instance-1-collector ScalingReplicaSet Scaled up replica set instance-1-collector-545ff6d759 to 1 deployment-controller logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:27 +0000 UTC Normal Pod instance-1-query-f6c5866cb-97gl9 Binding Scheduled Successfully assigned kuttl-test-improved-koi/instance-1-query-f6c5866cb-97gl9 to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:27 +0000 UTC Normal ReplicaSet.apps instance-1-query-f6c5866cb SuccessfulCreate Created pod: instance-1-query-f6c5866cb-97gl9 replicaset-controller logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:27 +0000 UTC Normal Deployment.apps instance-1-query ScalingReplicaSet Scaled up replica set instance-1-query-f6c5866cb to 1 deployment-controller logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:28 +0000 UTC Normal Pod instance-1-collector-545ff6d759-mpgt5 AddedInterface Add eth0 [10.129.2.41/23] from ovn-kubernetes multus logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:28 +0000 UTC Normal Pod instance-1-collector-545ff6d759-mpgt5.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:28 +0000 UTC Normal Pod instance-1-collector-545ff6d759-mpgt5.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:28 +0000 UTC Normal Pod instance-1-collector-545ff6d759-mpgt5.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:28 +0000 UTC Normal Pod instance-1-query-f6c5866cb-97gl9 AddedInterface Add eth0 [10.131.0.53/23] from ovn-kubernetes multus logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:28 +0000 UTC Normal Pod instance-1-query-f6c5866cb-97gl9.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:28 +0000 UTC Normal Pod instance-1-query-f6c5866cb-97gl9.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:28 +0000 UTC Normal Pod instance-1-query-f6c5866cb-97gl9.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:28 +0000 UTC Normal Pod instance-1-query-f6c5866cb-97gl9.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:28 +0000 UTC Normal Pod instance-1-query-f6c5866cb-97gl9.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:28 +0000 UTC Normal Pod instance-1-query-f6c5866cb-97gl9.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:28 +0000 UTC Normal Pod instance-1-query-f6c5866cb-97gl9.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:28 +0000 UTC Normal Pod instance-1-query-f6c5866cb-97gl9.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:28 +0000 UTC Normal Pod instance-1-query-f6c5866cb-97gl9.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:42 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:42 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:42 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:57 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:57 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod instance-1-collector-545ff6d759-mpgt5 horizontal-pod-autoscaler logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:03:57 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:04:12 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod instance-1-collector-545ff6d759-mpgt5 horizontal-pod-autoscaler logger.go:42: 18:04:21 | es-multiinstance | 2024-09-30 18:04:12 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod instance-1-collector-545ff6d759-mpgt5 horizontal-pod-autoscaler logger.go:42: 18:04:21 | es-multiinstance | Deleting namespace: kuttl-test-improved-koi === CONT kuttl/harness/es-simple-prod logger.go:42: 18:04:29 | es-simple-prod | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 18:04:29 | es-simple-prod | Creating namespace: kuttl-test-elegant-dragon logger.go:42: 18:04:29 | es-simple-prod | es-simple-prod events from ns kuttl-test-elegant-dragon: logger.go:42: 18:04:29 | es-simple-prod | Deleting namespace: kuttl-test-elegant-dragon === CONT kuttl/harness/es-rollover-autoprov logger.go:42: 18:04:35 | es-rollover-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 18:04:35 | es-rollover-autoprov | Creating namespace: kuttl-test-premium-walleye logger.go:42: 18:04:35 | es-rollover-autoprov/1-install | starting test step 1-install logger.go:42: 18:04:35 | es-rollover-autoprov/1-install | Jaeger:kuttl-test-premium-walleye/my-jaeger created logger.go:42: 18:05:13 | es-rollover-autoprov/1-install | test step completed 1-install logger.go:42: 18:05:13 | es-rollover-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 18:05:13 | es-rollover-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 18:05:15 | es-rollover-autoprov/2-report-spans | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 18:05:23 | es-rollover-autoprov/2-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 18:05:23 | es-rollover-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 18:05:23 | es-rollover-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 18:05:48 | es-rollover-autoprov/2-report-spans | test step completed 2-report-spans logger.go:42: 18:05:48 | es-rollover-autoprov/3-check-indices | starting test step 3-check-indices logger.go:42: 18:05:48 | es-rollover-autoprov/3-check-indices | Job:kuttl-test-premium-walleye/00-check-indices created logger.go:42: 18:05:52 | es-rollover-autoprov/3-check-indices | test step completed 3-check-indices logger.go:42: 18:05:52 | es-rollover-autoprov/4-check-indices | starting test step 4-check-indices logger.go:42: 18:05:52 | es-rollover-autoprov/4-check-indices | Job:kuttl-test-premium-walleye/01-check-indices created logger.go:42: 18:05:56 | es-rollover-autoprov/4-check-indices | test step completed 4-check-indices logger.go:42: 18:05:56 | es-rollover-autoprov/5-install | starting test step 5-install logger.go:42: 18:05:56 | es-rollover-autoprov/5-install | Jaeger:kuttl-test-premium-walleye/my-jaeger updated logger.go:42: 18:06:09 | es-rollover-autoprov/5-install | test step completed 5-install logger.go:42: 18:06:09 | es-rollover-autoprov/6-report-spans | starting test step 6-report-spans logger.go:42: 18:06:09 | es-rollover-autoprov/6-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 18:06:17 | es-rollover-autoprov/6-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=02 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-02-job.yaml] logger.go:42: 18:06:18 | es-rollover-autoprov/6-report-spans | running command: [sh -c kubectl apply -f report-span-02-job.yaml -n $NAMESPACE] logger.go:42: 18:06:18 | es-rollover-autoprov/6-report-spans | job.batch/02-report-span created logger.go:42: 18:06:42 | es-rollover-autoprov/6-report-spans | test step completed 6-report-spans logger.go:42: 18:06:42 | es-rollover-autoprov/7-check-indices | starting test step 7-check-indices logger.go:42: 18:06:42 | es-rollover-autoprov/7-check-indices | Job:kuttl-test-premium-walleye/02-check-indices created logger.go:42: 18:06:46 | es-rollover-autoprov/7-check-indices | test step completed 7-check-indices logger.go:42: 18:06:46 | es-rollover-autoprov/8-check-indices | starting test step 8-check-indices logger.go:42: 18:06:47 | es-rollover-autoprov/8-check-indices | Job:kuttl-test-premium-walleye/03-check-indices created logger.go:42: 18:06:51 | es-rollover-autoprov/8-check-indices | test step completed 8-check-indices logger.go:42: 18:06:51 | es-rollover-autoprov/9-check-indices | starting test step 9-check-indices logger.go:42: 18:06:51 | es-rollover-autoprov/9-check-indices | Job:kuttl-test-premium-walleye/04-check-indices created logger.go:42: 18:06:55 | es-rollover-autoprov/9-check-indices | test step completed 9-check-indices logger.go:42: 18:06:55 | es-rollover-autoprov/10-report-spans | starting test step 10-report-spans logger.go:42: 18:06:55 | es-rollover-autoprov/10-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 18:07:06 | es-rollover-autoprov/10-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=03 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-03-job.yaml] logger.go:42: 18:07:06 | es-rollover-autoprov/10-report-spans | running command: [sh -c kubectl apply -f report-span-03-job.yaml -n $NAMESPACE] logger.go:42: 18:07:06 | es-rollover-autoprov/10-report-spans | job.batch/03-report-span created logger.go:42: 18:07:30 | es-rollover-autoprov/10-report-spans | test step completed 10-report-spans logger.go:42: 18:07:30 | es-rollover-autoprov/11-check-indices | starting test step 11-check-indices logger.go:42: 18:07:30 | es-rollover-autoprov/11-check-indices | running command: [sh -c go run ../../../../cmd-utils/wait-cronjob/main.go --cronjob my-jaeger-es-rollover --namespace $NAMESPACE] logger.go:42: 18:07:41 | es-rollover-autoprov/11-check-indices | time="2024-09-30T18:07:41Z" level=debug msg="Checking if the my-jaeger-es-rollover CronJob exists" logger.go:42: 18:07:41 | es-rollover-autoprov/11-check-indices | time="2024-09-30T18:07:41Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 18:07:41 | es-rollover-autoprov/11-check-indices | time="2024-09-30T18:07:41Z" level=info msg="Cronjob my-jaeger-es-rollover found successfully" logger.go:42: 18:07:41 | es-rollover-autoprov/11-check-indices | time="2024-09-30T18:07:41Z" level=debug msg="Waiting for the next scheduled job from my-jaeger-es-rollover cronjob" logger.go:42: 18:07:41 | es-rollover-autoprov/11-check-indices | time="2024-09-30T18:07:41Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 18:07:51 | es-rollover-autoprov/11-check-indices | time="2024-09-30T18:07:51Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 18:08:01 | es-rollover-autoprov/11-check-indices | time="2024-09-30T18:08:01Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 18:08:11 | es-rollover-autoprov/11-check-indices | time="2024-09-30T18:08:11Z" level=info msg="Job of owner my-jaeger-es-rollover succeeded after my-jaeger-es-rollover 30.038257527s" logger.go:42: 18:08:11 | es-rollover-autoprov/11-check-indices | Job:kuttl-test-premium-walleye/05-check-indices created logger.go:42: 18:08:15 | es-rollover-autoprov/11-check-indices | test step completed 11-check-indices logger.go:42: 18:08:15 | es-rollover-autoprov/12-check-indices | starting test step 12-check-indices logger.go:42: 18:08:16 | es-rollover-autoprov/12-check-indices | Job:kuttl-test-premium-walleye/06-check-indices created logger.go:42: 18:08:20 | es-rollover-autoprov/12-check-indices | test step completed 12-check-indices logger.go:42: 18:08:20 | es-rollover-autoprov | es-rollover-autoprov events from ns kuttl-test-premium-walleye: logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:04:43 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestpremiumwalleyemyjaeger-1-66466544f6 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestpremiumwalleyemyjaeger-1-664665wgnc7 replicaset-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:04:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpremiumwalleyemyjaeger-1-664665wgnc7 Binding Scheduled Successfully assigned kuttl-test-premium-walleye/elasticsearch-cdm-kuttltestpremiumwalleyemyjaeger-1-664665wgnc7 to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:04:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpremiumwalleyemyjaeger-1-664665wgnc7 AddedInterface Add eth0 [10.128.2.30/23] from ovn-kubernetes multus logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:04:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpremiumwalleyemyjaeger-1-664665wgnc7.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:89629964e44058bafc90393a235b08c7c974f05513aecbfe7229134da732f2b5" already present on machine kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:04:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpremiumwalleyemyjaeger-1-664665wgnc7.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:04:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpremiumwalleyemyjaeger-1-664665wgnc7.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:04:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpremiumwalleyemyjaeger-1-664665wgnc7.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:a8f53dec42a46c5bf8ac7f2888848c01e70f405d27211e2a269730c888929faf" already present on machine kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:04:43 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestpremiumwalleyemyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestpremiumwalleyemyjaeger-1-66466544f6 to 1 deployment-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:04:44 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpremiumwalleyemyjaeger-1-664665wgnc7.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:04:44 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpremiumwalleyemyjaeger-1-664665wgnc7.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:04:53 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestpremiumwalleyemyjaeger-1-664665wgnc7.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:04:58 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestpremiumwalleyemyjaeger-1-664665wgnc7.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:10 +0000 UTC Normal Pod my-jaeger-collector-6d75985dfb-x2zgb Binding Scheduled Successfully assigned kuttl-test-premium-walleye/my-jaeger-collector-6d75985dfb-x2zgb to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:10 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-6d75985dfb SuccessfulCreate Created pod: my-jaeger-collector-6d75985dfb-x2zgb replicaset-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:10 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-6d75985dfb to 1 deployment-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:10 +0000 UTC Normal Pod my-jaeger-query-dd46cd58d-xt6zt Binding Scheduled Successfully assigned kuttl-test-premium-walleye/my-jaeger-query-dd46cd58d-xt6zt to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:10 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-dd46cd58d SuccessfulCreate Created pod: my-jaeger-query-dd46cd58d-xt6zt replicaset-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:10 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-dd46cd58d to 1 deployment-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:11 +0000 UTC Warning Pod my-jaeger-collector-6d75985dfb-x2zgb FailedMount MountVolume.SetUp failed for volume "certs" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:11 +0000 UTC Warning Pod my-jaeger-collector-6d75985dfb-x2zgb FailedMount MountVolume.SetUp failed for volume "my-jaeger-collector-tls-config-volume" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:11 +0000 UTC Warning Pod my-jaeger-collector-6d75985dfb-x2zgb FailedMount MountVolume.SetUp failed for volume "my-jaeger-sampling-configuration-volume" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:11 +0000 UTC Warning Pod my-jaeger-collector-6d75985dfb-x2zgb FailedMount MountVolume.SetUp failed for volume "kube-api-access-hf6sh" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:11 +0000 UTC Warning Pod my-jaeger-query-dd46cd58d-xt6zt FailedMount MountVolume.SetUp failed for volume "my-jaeger-trusted-ca" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:11 +0000 UTC Warning Pod my-jaeger-query-dd46cd58d-xt6zt FailedMount MountVolume.SetUp failed for volume "my-jaeger-ui-oauth-proxy-tls" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:11 +0000 UTC Warning Pod my-jaeger-query-dd46cd58d-xt6zt FailedMount MountVolume.SetUp failed for volume "my-jaeger-ui-configuration-volume" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:11 +0000 UTC Warning Pod my-jaeger-query-dd46cd58d-xt6zt FailedMount MountVolume.SetUp failed for volume "certs" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:11 +0000 UTC Warning Pod my-jaeger-query-dd46cd58d-xt6zt FailedMount MountVolume.SetUp failed for volume "kube-api-access-lwqt2" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:12 +0000 UTC Normal Pod my-jaeger-collector-6d75985dfb-x2zgb AddedInterface Add eth0 [10.129.2.43/23] from ovn-kubernetes multus logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:12 +0000 UTC Normal Pod my-jaeger-collector-6d75985dfb-x2zgb.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:12 +0000 UTC Normal Pod my-jaeger-collector-6d75985dfb-x2zgb.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:12 +0000 UTC Normal Pod my-jaeger-collector-6d75985dfb-x2zgb.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:12 +0000 UTC Normal Pod my-jaeger-query-dd46cd58d-xt6zt AddedInterface Add eth0 [10.131.0.56/23] from ovn-kubernetes multus logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:12 +0000 UTC Normal Pod my-jaeger-query-dd46cd58d-xt6zt.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:12 +0000 UTC Normal Pod my-jaeger-query-dd46cd58d-xt6zt.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:12 +0000 UTC Normal Pod my-jaeger-query-dd46cd58d-xt6zt.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:12 +0000 UTC Normal Pod my-jaeger-query-dd46cd58d-xt6zt.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:12 +0000 UTC Normal Pod my-jaeger-query-dd46cd58d-xt6zt.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:12 +0000 UTC Normal Pod my-jaeger-query-dd46cd58d-xt6zt.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:12 +0000 UTC Normal Pod my-jaeger-query-dd46cd58d-xt6zt.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:12 +0000 UTC Normal Pod my-jaeger-query-dd46cd58d-xt6zt.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:12 +0000 UTC Normal Pod my-jaeger-query-dd46cd58d-xt6zt.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:19 +0000 UTC Normal Pod my-jaeger-query-dd46cd58d-xt6zt.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:19 +0000 UTC Normal Pod my-jaeger-query-dd46cd58d-xt6zt.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:19 +0000 UTC Normal Pod my-jaeger-query-dd46cd58d-xt6zt.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:19 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-dd46cd58d SuccessfulDelete Deleted pod: my-jaeger-query-dd46cd58d-xt6zt replicaset-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:19 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-dd46cd58d to 0 from 1 deployment-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:20 +0000 UTC Normal Pod my-jaeger-query-5b9d5979d9-rz8nn Binding Scheduled Successfully assigned kuttl-test-premium-walleye/my-jaeger-query-5b9d5979d9-rz8nn to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:20 +0000 UTC Normal Pod my-jaeger-query-5b9d5979d9-rz8nn AddedInterface Add eth0 [10.131.0.57/23] from ovn-kubernetes multus logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:20 +0000 UTC Normal Pod my-jaeger-query-5b9d5979d9-rz8nn.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:20 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-5b9d5979d9 SuccessfulCreate Created pod: my-jaeger-query-5b9d5979d9-rz8nn replicaset-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:20 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-5b9d5979d9 to 1 deployment-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:21 +0000 UTC Normal Pod my-jaeger-query-5b9d5979d9-rz8nn.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:21 +0000 UTC Normal Pod my-jaeger-query-5b9d5979d9-rz8nn.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:21 +0000 UTC Normal Pod my-jaeger-query-5b9d5979d9-rz8nn.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:21 +0000 UTC Normal Pod my-jaeger-query-5b9d5979d9-rz8nn.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:21 +0000 UTC Normal Pod my-jaeger-query-5b9d5979d9-rz8nn.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:21 +0000 UTC Normal Pod my-jaeger-query-5b9d5979d9-rz8nn.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:21 +0000 UTC Normal Pod my-jaeger-query-5b9d5979d9-rz8nn.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:21 +0000 UTC Normal Pod my-jaeger-query-5b9d5979d9-rz8nn.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:23 +0000 UTC Normal Pod 00-report-span-lsk5t Binding Scheduled Successfully assigned kuttl-test-premium-walleye/00-report-span-lsk5t to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:23 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-lsk5t job-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:24 +0000 UTC Normal Pod 00-report-span-lsk5t AddedInterface Add eth0 [10.129.2.44/23] from ovn-kubernetes multus logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:24 +0000 UTC Normal Pod 00-report-span-lsk5t.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:24 +0000 UTC Normal Pod 00-report-span-lsk5t.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 296ms (296ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:24 +0000 UTC Normal Pod 00-report-span-lsk5t.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:25 +0000 UTC Normal Pod 00-report-span-lsk5t.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:25 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:25 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:25 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:40 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:40 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-6d75985dfb-x2zgb horizontal-pod-autoscaler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:40 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:47 +0000 UTC Normal Job.batch 00-report-span Completed Job completed job-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:48 +0000 UTC Normal Pod 00-check-indices-d45xq Binding Scheduled Successfully assigned kuttl-test-premium-walleye/00-check-indices-d45xq to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:48 +0000 UTC Normal Job.batch 00-check-indices SuccessfulCreate Created pod: 00-check-indices-d45xq job-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:49 +0000 UTC Normal Pod 00-check-indices-d45xq AddedInterface Add eth0 [10.129.2.45/23] from ovn-kubernetes multus logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:49 +0000 UTC Normal Pod 00-check-indices-d45xq.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:49 +0000 UTC Normal Pod 00-check-indices-d45xq.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 199ms (199ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:49 +0000 UTC Normal Pod 00-check-indices-d45xq.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:49 +0000 UTC Normal Pod 00-check-indices-d45xq.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:52 +0000 UTC Normal Job.batch 00-check-indices Completed Job completed job-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:52 +0000 UTC Normal Pod 01-check-indices-zqmlf Binding Scheduled Successfully assigned kuttl-test-premium-walleye/01-check-indices-zqmlf to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:52 +0000 UTC Normal Job.batch 01-check-indices SuccessfulCreate Created pod: 01-check-indices-zqmlf job-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:53 +0000 UTC Normal Pod 01-check-indices-zqmlf AddedInterface Add eth0 [10.129.2.46/23] from ovn-kubernetes multus logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:53 +0000 UTC Normal Pod 01-check-indices-zqmlf.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:53 +0000 UTC Normal Pod 01-check-indices-zqmlf.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 195ms (196ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:53 +0000 UTC Normal Pod 01-check-indices-zqmlf.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:53 +0000 UTC Normal Pod 01-check-indices-zqmlf.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:55 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-6d75985dfb-x2zgb horizontal-pod-autoscaler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:55 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-6d75985dfb-x2zgb horizontal-pod-autoscaler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:56 +0000 UTC Normal Job.batch 01-check-indices Completed Job completed job-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:58 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-rsxd9 Binding Scheduled Successfully assigned kuttl-test-premium-walleye/my-jaeger-es-rollover-create-mapping-rsxd9 to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:58 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-rsxd9 AddedInterface Add eth0 [10.129.2.47/23] from ovn-kubernetes multus logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:58 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-rsxd9.spec.containers{my-jaeger-es-rollover-create-mapping} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:b3897239050ab12ae12c28bb51d955438334aad957dceb83a53cf2e8a466b69b" kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:05:58 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping SuccessfulCreate Created pod: my-jaeger-es-rollover-create-mapping-rsxd9 job-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:05 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-rsxd9.spec.containers{my-jaeger-es-rollover-create-mapping} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:b3897239050ab12ae12c28bb51d955438334aad957dceb83a53cf2e8a466b69b" in 7.055s (7.055s including waiting). Image size: 111568812 bytes. kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:05 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-rsxd9.spec.containers{my-jaeger-es-rollover-create-mapping} Created Created container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:06 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-rsxd9.spec.containers{my-jaeger-es-rollover-create-mapping} Started Started container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:09 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping Completed Job completed job-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:10 +0000 UTC Normal Pod my-jaeger-collector-6d75985dfb-x2zgb.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:10 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-6d75985dfb SuccessfulDelete Deleted pod: my-jaeger-collector-6d75985dfb-x2zgb replicaset-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:10 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-6d75985dfb to 0 from 1 deployment-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:10 +0000 UTC Normal Pod my-jaeger-query-5b9d5979d9-rz8nn.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:10 +0000 UTC Normal Pod my-jaeger-query-5b9d5979d9-rz8nn.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:10 +0000 UTC Normal Pod my-jaeger-query-5b9d5979d9-rz8nn.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:10 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-5b9d5979d9 SuccessfulDelete Deleted pod: my-jaeger-query-5b9d5979d9-rz8nn replicaset-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:10 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-5b9d5979d9 to 0 from 1 deployment-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:11 +0000 UTC Normal Pod my-jaeger-collector-69c95b747c-nwd9v Binding Scheduled Successfully assigned kuttl-test-premium-walleye/my-jaeger-collector-69c95b747c-nwd9v to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:11 +0000 UTC Normal Pod my-jaeger-collector-69c95b747c-nwd9v AddedInterface Add eth0 [10.131.0.58/23] from ovn-kubernetes multus logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:11 +0000 UTC Normal Pod my-jaeger-collector-69c95b747c-nwd9v.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:11 +0000 UTC Normal Pod my-jaeger-collector-69c95b747c-nwd9v.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:11 +0000 UTC Normal Pod my-jaeger-collector-69c95b747c-nwd9v.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:11 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-69c95b747c SuccessfulCreate Created pod: my-jaeger-collector-69c95b747c-nwd9v replicaset-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:11 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-69c95b747c to 1 deployment-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:11 +0000 UTC Normal Pod my-jaeger-query-5bb69b5c7b-4424v Binding Scheduled Successfully assigned kuttl-test-premium-walleye/my-jaeger-query-5bb69b5c7b-4424v to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:11 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-5bb69b5c7b SuccessfulCreate Created pod: my-jaeger-query-5bb69b5c7b-4424v replicaset-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:11 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-5bb69b5c7b to 1 deployment-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:12 +0000 UTC Normal Pod my-jaeger-query-5bb69b5c7b-4424v AddedInterface Add eth0 [10.131.0.59/23] from ovn-kubernetes multus logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:12 +0000 UTC Normal Pod my-jaeger-query-5bb69b5c7b-4424v.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:12 +0000 UTC Normal Pod my-jaeger-query-5bb69b5c7b-4424v.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:12 +0000 UTC Normal Pod my-jaeger-query-5bb69b5c7b-4424v.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:12 +0000 UTC Normal Pod my-jaeger-query-5bb69b5c7b-4424v.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:12 +0000 UTC Normal Pod my-jaeger-query-5bb69b5c7b-4424v.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:12 +0000 UTC Normal Pod my-jaeger-query-5bb69b5c7b-4424v.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:12 +0000 UTC Normal Pod my-jaeger-query-5bb69b5c7b-4424v.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:12 +0000 UTC Normal Pod my-jaeger-query-5bb69b5c7b-4424v.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:12 +0000 UTC Normal Pod my-jaeger-query-5bb69b5c7b-4424v.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:18 +0000 UTC Normal Pod 02-report-span-gpz8f Binding Scheduled Successfully assigned kuttl-test-premium-walleye/02-report-span-gpz8f to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:18 +0000 UTC Normal Pod 02-report-span-gpz8f AddedInterface Add eth0 [10.129.2.48/23] from ovn-kubernetes multus logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:18 +0000 UTC Normal Pod 02-report-span-gpz8f.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:18 +0000 UTC Normal Job.batch 02-report-span SuccessfulCreate Created pod: 02-report-span-gpz8f job-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:19 +0000 UTC Normal Pod 02-report-span-gpz8f.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 197ms (197ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:19 +0000 UTC Normal Pod 02-report-span-gpz8f.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:19 +0000 UTC Normal Pod 02-report-span-gpz8f.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:40 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-69c95b747c-nwd9v horizontal-pod-autoscaler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:42 +0000 UTC Normal Pod 02-check-indices-ml46f Binding Scheduled Successfully assigned kuttl-test-premium-walleye/02-check-indices-ml46f to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:42 +0000 UTC Normal Job.batch 02-check-indices SuccessfulCreate Created pod: 02-check-indices-ml46f job-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:42 +0000 UTC Normal Job.batch 02-report-span Completed Job completed job-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:43 +0000 UTC Normal Pod 02-check-indices-ml46f AddedInterface Add eth0 [10.129.2.49/23] from ovn-kubernetes multus logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:43 +0000 UTC Normal Pod 02-check-indices-ml46f.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:43 +0000 UTC Normal Pod 02-check-indices-ml46f.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 233ms (233ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:43 +0000 UTC Normal Pod 02-check-indices-ml46f.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:43 +0000 UTC Normal Pod 02-check-indices-ml46f.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:46 +0000 UTC Normal Job.batch 02-check-indices Completed Job completed job-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:47 +0000 UTC Normal Pod 03-check-indices-p66sv Binding Scheduled Successfully assigned kuttl-test-premium-walleye/03-check-indices-p66sv to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:47 +0000 UTC Normal Pod 03-check-indices-p66sv AddedInterface Add eth0 [10.129.2.50/23] from ovn-kubernetes multus logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:47 +0000 UTC Normal Pod 03-check-indices-p66sv.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:47 +0000 UTC Normal Pod 03-check-indices-p66sv.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 181ms (181ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:47 +0000 UTC Normal Pod 03-check-indices-p66sv.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:47 +0000 UTC Normal Pod 03-check-indices-p66sv.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:47 +0000 UTC Normal Job.batch 03-check-indices SuccessfulCreate Created pod: 03-check-indices-p66sv job-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:50 +0000 UTC Normal Job.batch 03-check-indices Completed Job completed job-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:51 +0000 UTC Normal Pod 04-check-indices-m49km Binding Scheduled Successfully assigned kuttl-test-premium-walleye/04-check-indices-m49km to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:51 +0000 UTC Normal Pod 04-check-indices-m49km AddedInterface Add eth0 [10.129.2.51/23] from ovn-kubernetes multus logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:51 +0000 UTC Normal Pod 04-check-indices-m49km.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:51 +0000 UTC Normal Job.batch 04-check-indices SuccessfulCreate Created pod: 04-check-indices-m49km job-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:52 +0000 UTC Normal Pod 04-check-indices-m49km.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 259ms (259ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:52 +0000 UTC Normal Pod 04-check-indices-m49km.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:52 +0000 UTC Normal Pod 04-check-indices-m49km.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:55 +0000 UTC Normal Job.batch 04-check-indices Completed Job completed job-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:55 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-69c95b747c-nwd9v horizontal-pod-autoscaler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:06:55 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-69c95b747c-nwd9v horizontal-pod-autoscaler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:07:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28795327-298mb Binding Scheduled Successfully assigned kuttl-test-premium-walleye/my-jaeger-es-lookback-28795327-298mb to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:07:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28795327-298mb AddedInterface Add eth0 [10.129.2.52/23] from ovn-kubernetes multus logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:07:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28795327-298mb.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:b3897239050ab12ae12c28bb51d955438334aad957dceb83a53cf2e8a466b69b" already present on machine kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:07:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28795327-298mb.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:07:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28795327-298mb.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:07:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28795327 SuccessfulCreate Created pod: my-jaeger-es-lookback-28795327-298mb job-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:07:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28795327 cronjob-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:07:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28795327-rdp4b Binding Scheduled Successfully assigned kuttl-test-premium-walleye/my-jaeger-es-rollover-28795327-rdp4b to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:07:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28795327-rdp4b AddedInterface Add eth0 [10.129.2.53/23] from ovn-kubernetes multus logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:07:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28795327-rdp4b.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:b3897239050ab12ae12c28bb51d955438334aad957dceb83a53cf2e8a466b69b" already present on machine kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:07:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28795327-rdp4b.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:07:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28795327 SuccessfulCreate Created pod: my-jaeger-es-rollover-28795327-rdp4b job-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:07:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28795327 cronjob-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:07:01 +0000 UTC Normal Pod my-jaeger-es-rollover-28795327-rdp4b.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:07:03 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28795327 Completed Job completed job-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:07:03 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28795327, status: Complete cronjob-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:07:03 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28795327 Completed Job completed job-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:07:03 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28795327, status: Complete cronjob-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:07:06 +0000 UTC Normal Pod 03-report-span-g7bt2 Binding Scheduled Successfully assigned kuttl-test-premium-walleye/03-report-span-g7bt2 to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:07:06 +0000 UTC Normal Job.batch 03-report-span SuccessfulCreate Created pod: 03-report-span-g7bt2 job-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:07:07 +0000 UTC Normal Pod 03-report-span-g7bt2 AddedInterface Add eth0 [10.129.2.54/23] from ovn-kubernetes multus logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:07:07 +0000 UTC Normal Pod 03-report-span-g7bt2.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:07:07 +0000 UTC Normal Pod 03-report-span-g7bt2.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 251ms (251ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:07:07 +0000 UTC Normal Pod 03-report-span-g7bt2.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:07:07 +0000 UTC Normal Pod 03-report-span-g7bt2.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:07:30 +0000 UTC Normal Job.batch 03-report-span Completed Job completed job-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28795328-slm7j Binding Scheduled Successfully assigned kuttl-test-premium-walleye/my-jaeger-es-lookback-28795328-slm7j to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28795328-slm7j AddedInterface Add eth0 [10.129.2.56/23] from ovn-kubernetes multus logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28795328-slm7j.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:b3897239050ab12ae12c28bb51d955438334aad957dceb83a53cf2e8a466b69b" already present on machine kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28795328-slm7j.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28795328 SuccessfulCreate Created pod: my-jaeger-es-lookback-28795328-slm7j job-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28795328 cronjob-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28795328-j2px2 Binding Scheduled Successfully assigned kuttl-test-premium-walleye/my-jaeger-es-rollover-28795328-j2px2 to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28795328-j2px2 AddedInterface Add eth0 [10.129.2.55/23] from ovn-kubernetes multus logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28795328-j2px2.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:b3897239050ab12ae12c28bb51d955438334aad957dceb83a53cf2e8a466b69b" already present on machine kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28795328-j2px2.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28795328-j2px2.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28795328 SuccessfulCreate Created pod: my-jaeger-es-rollover-28795328-j2px2 job-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28795328 cronjob-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:01 +0000 UTC Normal Pod my-jaeger-es-lookback-28795328-slm7j.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:03 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28795328 Completed Job completed job-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:03 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28795328, status: Complete cronjob-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:03 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28795328 Completed Job completed job-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:03 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28795328, status: Complete cronjob-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:11 +0000 UTC Normal Pod 05-check-indices-ppzg2 Binding Scheduled Successfully assigned kuttl-test-premium-walleye/05-check-indices-ppzg2 to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:11 +0000 UTC Normal Job.batch 05-check-indices SuccessfulCreate Created pod: 05-check-indices-ppzg2 job-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:12 +0000 UTC Normal Pod 05-check-indices-ppzg2 AddedInterface Add eth0 [10.129.2.57/23] from ovn-kubernetes multus logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:12 +0000 UTC Normal Pod 05-check-indices-ppzg2.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:13 +0000 UTC Normal Pod 05-check-indices-ppzg2.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 358ms (358ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:13 +0000 UTC Normal Pod 05-check-indices-ppzg2.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:13 +0000 UTC Normal Pod 05-check-indices-ppzg2.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:15 +0000 UTC Normal Job.batch 05-check-indices Completed Job completed job-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:16 +0000 UTC Normal Pod 06-check-indices-s5g7l Binding Scheduled Successfully assigned kuttl-test-premium-walleye/06-check-indices-s5g7l to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:16 +0000 UTC Normal Pod 06-check-indices-s5g7l AddedInterface Add eth0 [10.129.2.58/23] from ovn-kubernetes multus logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:16 +0000 UTC Normal Pod 06-check-indices-s5g7l.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:16 +0000 UTC Normal Pod 06-check-indices-s5g7l.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 226ms (226ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:16 +0000 UTC Normal Job.batch 06-check-indices SuccessfulCreate Created pod: 06-check-indices-s5g7l job-controller logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:17 +0000 UTC Normal Pod 06-check-indices-s5g7l.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:17 +0000 UTC Normal Pod 06-check-indices-s5g7l.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 18:08:20 | es-rollover-autoprov | 2024-09-30 18:08:19 +0000 UTC Normal Job.batch 06-check-indices Completed Job completed job-controller logger.go:42: 18:08:20 | es-rollover-autoprov | Deleting namespace: kuttl-test-premium-walleye === CONT kuttl/harness/es-increasing-replicas logger.go:42: 18:08:27 | es-increasing-replicas | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 18:08:27 | es-increasing-replicas | Ignoring check-es-nodes.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 18:08:27 | es-increasing-replicas | Ignoring openshift-check-es-nodes.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 18:08:27 | es-increasing-replicas | Creating namespace: kuttl-test-honest-termite logger.go:42: 18:08:27 | es-increasing-replicas/1-install | starting test step 1-install logger.go:42: 18:08:27 | es-increasing-replicas/1-install | Jaeger:kuttl-test-honest-termite/simple-prod created logger.go:42: 18:09:04 | es-increasing-replicas/1-install | test step completed 1-install logger.go:42: 18:09:04 | es-increasing-replicas/2-install | starting test step 2-install logger.go:42: 18:09:04 | es-increasing-replicas/2-install | Jaeger:kuttl-test-honest-termite/simple-prod updated logger.go:42: 18:09:17 | es-increasing-replicas/2-install | test step completed 2-install logger.go:42: 18:09:17 | es-increasing-replicas/3-smoke-test | starting test step 3-smoke-test logger.go:42: 18:09:17 | es-increasing-replicas/3-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 18:09:19 | es-increasing-replicas/3-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 18:09:25 | es-increasing-replicas/3-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 18:09:26 | es-increasing-replicas/3-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 18:09:26 | es-increasing-replicas/3-smoke-test | job.batch/report-span created logger.go:42: 18:09:26 | es-increasing-replicas/3-smoke-test | job.batch/check-span created logger.go:42: 18:09:38 | es-increasing-replicas/3-smoke-test | test step completed 3-smoke-test logger.go:42: 18:09:38 | es-increasing-replicas/4-install | starting test step 4-install logger.go:42: 18:09:39 | es-increasing-replicas/4-install | Jaeger:kuttl-test-honest-termite/simple-prod updated logger.go:42: 18:09:39 | es-increasing-replicas/4-install | test step completed 4-install logger.go:42: 18:09:39 | es-increasing-replicas/5-check-es-nodes | starting test step 5-check-es-nodes logger.go:42: 18:09:39 | es-increasing-replicas/5-check-es-nodes | running command: [sh -c ./check-es-nodes.sh $NAMESPACE] logger.go:42: 18:09:39 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 18:09:39 | es-increasing-replicas/5-check-es-nodes | false logger.go:42: 18:09:39 | es-increasing-replicas/5-check-es-nodes | Error: no matches found logger.go:42: 18:09:44 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 18:09:44 | es-increasing-replicas/5-check-es-nodes | true logger.go:42: 18:09:44 | es-increasing-replicas/5-check-es-nodes | test step completed 5-check-es-nodes logger.go:42: 18:09:44 | es-increasing-replicas | es-increasing-replicas events from ns kuttl-test-honest-termite: logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:08:34 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthonesttermitesimpleprod-1-5dbd54m95t Binding Scheduled Successfully assigned kuttl-test-honest-termite/elasticsearch-cdm-kuttltesthonesttermitesimpleprod-1-5dbd54m95t to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:08:34 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltesthonesttermitesimpleprod-1-5dbd56dd78 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltesthonesttermitesimpleprod-1-5dbd54m95t replicaset-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:08:34 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltesthonesttermitesimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltesthonesttermitesimpleprod-1-5dbd56dd78 to 1 deployment-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:08:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthonesttermitesimpleprod-1-5dbd54m95t AddedInterface Add eth0 [10.128.2.31/23] from ovn-kubernetes multus logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:08:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthonesttermitesimpleprod-1-5dbd54m95t.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:89629964e44058bafc90393a235b08c7c974f05513aecbfe7229134da732f2b5" already present on machine kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:08:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthonesttermitesimpleprod-1-5dbd54m95t.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:08:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthonesttermitesimpleprod-1-5dbd54m95t.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:08:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthonesttermitesimpleprod-1-5dbd54m95t.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:a8f53dec42a46c5bf8ac7f2888848c01e70f405d27211e2a269730c888929faf" already present on machine kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:08:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthonesttermitesimpleprod-1-5dbd54m95t.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:08:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthonesttermitesimpleprod-1-5dbd54m95t.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:08:50 +0000 UTC Warning Pod elasticsearch-cdm-kuttltesthonesttermitesimpleprod-1-5dbd54m95t.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:01 +0000 UTC Normal Pod simple-prod-collector-759d697875-wx9pp Binding Scheduled Successfully assigned kuttl-test-honest-termite/simple-prod-collector-759d697875-wx9pp to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:01 +0000 UTC Normal Pod simple-prod-collector-759d697875-wx9pp AddedInterface Add eth0 [10.129.2.59/23] from ovn-kubernetes multus logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:01 +0000 UTC Normal Pod simple-prod-collector-759d697875-wx9pp.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:01 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-759d697875 SuccessfulCreate Created pod: simple-prod-collector-759d697875-wx9pp replicaset-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:01 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-759d697875 to 1 deployment-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:01 +0000 UTC Normal Pod simple-prod-query-847974b65-28m9h Binding Scheduled Successfully assigned kuttl-test-honest-termite/simple-prod-query-847974b65-28m9h to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:01 +0000 UTC Normal Pod simple-prod-query-847974b65-28m9h AddedInterface Add eth0 [10.131.0.60/23] from ovn-kubernetes multus logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:01 +0000 UTC Normal Pod simple-prod-query-847974b65-28m9h.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:01 +0000 UTC Normal ReplicaSet.apps simple-prod-query-847974b65 SuccessfulCreate Created pod: simple-prod-query-847974b65-28m9h replicaset-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:01 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-847974b65 to 1 deployment-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:02 +0000 UTC Normal Pod simple-prod-collector-759d697875-wx9pp.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:02 +0000 UTC Normal Pod simple-prod-collector-759d697875-wx9pp.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:02 +0000 UTC Normal Pod simple-prod-query-847974b65-28m9h.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:02 +0000 UTC Normal Pod simple-prod-query-847974b65-28m9h.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:02 +0000 UTC Normal Pod simple-prod-query-847974b65-28m9h.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:02 +0000 UTC Normal Pod simple-prod-query-847974b65-28m9h.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:02 +0000 UTC Normal Pod simple-prod-query-847974b65-28m9h.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:02 +0000 UTC Normal Pod simple-prod-query-847974b65-28m9h.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:02 +0000 UTC Normal Pod simple-prod-query-847974b65-28m9h.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:02 +0000 UTC Normal Pod simple-prod-query-847974b65-28m9h.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:05 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-759d697875 to 2 from 1 deployment-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:06 +0000 UTC Normal Pod simple-prod-collector-759d697875-p5mvv Binding Scheduled Successfully assigned kuttl-test-honest-termite/simple-prod-collector-759d697875-p5mvv to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:06 +0000 UTC Normal Pod simple-prod-collector-759d697875-p5mvv AddedInterface Add eth0 [10.131.0.61/23] from ovn-kubernetes multus logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:06 +0000 UTC Normal Pod simple-prod-collector-759d697875-p5mvv.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:06 +0000 UTC Normal Pod simple-prod-collector-759d697875-p5mvv.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:06 +0000 UTC Normal Pod simple-prod-collector-759d697875-p5mvv.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:06 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-759d697875 SuccessfulCreate Created pod: simple-prod-collector-759d697875-p5mvv replicaset-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:06 +0000 UTC Normal Pod simple-prod-query-847974b65-hs6ds Binding Scheduled Successfully assigned kuttl-test-honest-termite/simple-prod-query-847974b65-hs6ds to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:06 +0000 UTC Normal ReplicaSet.apps simple-prod-query-847974b65 SuccessfulCreate Created pod: simple-prod-query-847974b65-hs6ds replicaset-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:06 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-847974b65 to 2 from 1 deployment-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:07 +0000 UTC Warning Pod simple-prod-query-847974b65-hs6ds FailedMount MountVolume.SetUp failed for volume "simple-prod-ui-oauth-proxy-tls" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:07 +0000 UTC Warning Pod simple-prod-query-847974b65-hs6ds FailedMount MountVolume.SetUp failed for volume "simple-prod-service-ca" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:08 +0000 UTC Normal Pod simple-prod-query-847974b65-hs6ds AddedInterface Add eth0 [10.129.2.60/23] from ovn-kubernetes multus logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:08 +0000 UTC Normal Pod simple-prod-query-847974b65-hs6ds.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:12 +0000 UTC Normal Pod simple-prod-query-847974b65-hs6ds.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" in 3.972s (3.972s including waiting). Image size: 192762590 bytes. kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:12 +0000 UTC Normal Pod simple-prod-query-847974b65-hs6ds.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:12 +0000 UTC Normal Pod simple-prod-query-847974b65-hs6ds.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:12 +0000 UTC Normal Pod simple-prod-query-847974b65-hs6ds.spec.containers{oauth-proxy} Pulling Pulling image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:15 +0000 UTC Normal Pod simple-prod-query-847974b65-hs6ds.spec.containers{oauth-proxy} Pulled Successfully pulled image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" in 3.476s (3.476s including waiting). Image size: 339954870 bytes. kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:15 +0000 UTC Normal Pod simple-prod-query-847974b65-hs6ds.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:15 +0000 UTC Normal Pod simple-prod-query-847974b65-hs6ds.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:15 +0000 UTC Normal Pod simple-prod-query-847974b65-hs6ds.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:16 +0000 UTC Normal Pod simple-prod-query-847974b65-hs6ds.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:16 +0000 UTC Normal Pod simple-prod-query-847974b65-hs6ds.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:21 +0000 UTC Normal Pod simple-prod-query-6c768db549-9jwvh Binding Scheduled Successfully assigned kuttl-test-honest-termite/simple-prod-query-6c768db549-9jwvh to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:21 +0000 UTC Normal Pod simple-prod-query-6c768db549-hgwd8 Binding Scheduled Successfully assigned kuttl-test-honest-termite/simple-prod-query-6c768db549-hgwd8 to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:21 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6c768db549 SuccessfulCreate Created pod: simple-prod-query-6c768db549-hgwd8 replicaset-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:21 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6c768db549 SuccessfulCreate Created pod: simple-prod-query-6c768db549-9jwvh replicaset-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:21 +0000 UTC Normal Pod simple-prod-query-847974b65-28m9h.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:21 +0000 UTC Normal Pod simple-prod-query-847974b65-28m9h.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:21 +0000 UTC Normal Pod simple-prod-query-847974b65-28m9h.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:21 +0000 UTC Normal Pod simple-prod-query-847974b65-hs6ds.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:21 +0000 UTC Normal Pod simple-prod-query-847974b65-hs6ds.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:21 +0000 UTC Normal Pod simple-prod-query-847974b65-hs6ds.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:21 +0000 UTC Normal ReplicaSet.apps simple-prod-query-847974b65 SuccessfulDelete Deleted pod: simple-prod-query-847974b65-hs6ds replicaset-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:21 +0000 UTC Normal ReplicaSet.apps simple-prod-query-847974b65 SuccessfulDelete Deleted pod: simple-prod-query-847974b65-28m9h replicaset-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:21 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-847974b65 to 0 from 2 deployment-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:21 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-6c768db549 to 2 deployment-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:22 +0000 UTC Normal Pod simple-prod-query-6c768db549-9jwvh AddedInterface Add eth0 [10.129.2.61/23] from ovn-kubernetes multus logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:22 +0000 UTC Normal Pod simple-prod-query-6c768db549-9jwvh.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:22 +0000 UTC Normal Pod simple-prod-query-6c768db549-9jwvh.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:22 +0000 UTC Normal Pod simple-prod-query-6c768db549-9jwvh.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:22 +0000 UTC Normal Pod simple-prod-query-6c768db549-9jwvh.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:22 +0000 UTC Normal Pod simple-prod-query-6c768db549-9jwvh.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:22 +0000 UTC Normal Pod simple-prod-query-6c768db549-9jwvh.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:22 +0000 UTC Normal Pod simple-prod-query-6c768db549-9jwvh.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:22 +0000 UTC Normal Pod simple-prod-query-6c768db549-9jwvh.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:22 +0000 UTC Normal Pod simple-prod-query-6c768db549-9jwvh.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:22 +0000 UTC Normal Pod simple-prod-query-6c768db549-hgwd8 AddedInterface Add eth0 [10.131.0.62/23] from ovn-kubernetes multus logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:22 +0000 UTC Normal Pod simple-prod-query-6c768db549-hgwd8.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:22 +0000 UTC Normal Pod simple-prod-query-6c768db549-hgwd8.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:22 +0000 UTC Normal Pod simple-prod-query-6c768db549-hgwd8.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:22 +0000 UTC Normal Pod simple-prod-query-6c768db549-hgwd8.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:22 +0000 UTC Normal Pod simple-prod-query-6c768db549-hgwd8.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:22 +0000 UTC Normal Pod simple-prod-query-6c768db549-hgwd8.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:22 +0000 UTC Normal Pod simple-prod-query-6c768db549-hgwd8.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:22 +0000 UTC Normal Pod simple-prod-query-6c768db549-hgwd8.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:22 +0000 UTC Normal Pod simple-prod-query-6c768db549-hgwd8.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:26 +0000 UTC Normal Pod check-span-r69cv Binding Scheduled Successfully assigned kuttl-test-honest-termite/check-span-r69cv to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:26 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-r69cv job-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:26 +0000 UTC Normal Pod report-span-j8qw2 Binding Scheduled Successfully assigned kuttl-test-honest-termite/report-span-j8qw2 to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:26 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-j8qw2 job-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:27 +0000 UTC Normal Pod check-span-r69cv AddedInterface Add eth0 [10.131.0.63/23] from ovn-kubernetes multus logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:27 +0000 UTC Normal Pod check-span-r69cv.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:27 +0000 UTC Normal Pod check-span-r69cv.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 194ms (194ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:27 +0000 UTC Normal Pod check-span-r69cv.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:27 +0000 UTC Normal Pod check-span-r69cv.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:27 +0000 UTC Normal Pod report-span-j8qw2 AddedInterface Add eth0 [10.129.2.62/23] from ovn-kubernetes multus logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:27 +0000 UTC Normal Pod report-span-j8qw2.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:27 +0000 UTC Normal Pod report-span-j8qw2.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 194ms (194ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:27 +0000 UTC Normal Pod report-span-j8qw2.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:27 +0000 UTC Normal Pod report-span-j8qw2.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:37 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthonesttermitesimpleprod-2-786bf4mpv6 Binding Scheduled Successfully assigned kuttl-test-honest-termite/elasticsearch-cdm-kuttltesthonesttermitesimpleprod-2-786bf4mpv6 to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:40 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltesthonesttermitesimpleprod-2-786bfb8d94 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltesthonesttermitesimpleprod-2-786bf4mpv6 replicaset-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:40 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltesthonesttermitesimpleprod-2 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltesthonesttermitesimpleprod-2-786bfb8d94 to 1 deployment-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:40 +0000 UTC Normal Pod simple-prod-collector-759d697875-p5mvv.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:40 +0000 UTC Normal Pod simple-prod-collector-759d697875-wx9pp.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:40 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-759d697875 SuccessfulDelete Deleted pod: simple-prod-collector-759d697875-wx9pp replicaset-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:40 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-759d697875 SuccessfulDelete Deleted pod: simple-prod-collector-759d697875-p5mvv replicaset-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:40 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled down replica set simple-prod-collector-759d697875 to 0 from 2 deployment-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:40 +0000 UTC Normal Pod simple-prod-query-6c768db549-9jwvh.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:40 +0000 UTC Normal Pod simple-prod-query-6c768db549-9jwvh.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:40 +0000 UTC Normal Pod simple-prod-query-6c768db549-9jwvh.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:40 +0000 UTC Normal Pod simple-prod-query-6c768db549-hgwd8.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:40 +0000 UTC Normal Pod simple-prod-query-6c768db549-hgwd8.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:40 +0000 UTC Normal Pod simple-prod-query-6c768db549-hgwd8.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:40 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6c768db549 SuccessfulDelete Deleted pod: simple-prod-query-6c768db549-9jwvh replicaset-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:40 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6c768db549 SuccessfulDelete Deleted pod: simple-prod-query-6c768db549-hgwd8 replicaset-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:40 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-6c768db549 to 0 from 2 deployment-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:41 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthonesttermitesimpleprod-2-786bf4mpv6 AddedInterface Add eth0 [10.129.2.63/23] from ovn-kubernetes multus logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:41 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthonesttermitesimpleprod-2-786bf4mpv6.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:89629964e44058bafc90393a235b08c7c974f05513aecbfe7229134da732f2b5" already present on machine kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:41 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthonesttermitesimpleprod-2-786bf4mpv6.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:41 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthonesttermitesimpleprod-2-786bf4mpv6.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:41 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthonesttermitesimpleprod-2-786bf4mpv6.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:a8f53dec42a46c5bf8ac7f2888848c01e70f405d27211e2a269730c888929faf" already present on machine kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:41 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthonesttermitesimpleprod-2-786bf4mpv6.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:41 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthonesttermitesimpleprod-2-786bf4mpv6.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:41 +0000 UTC Normal Pod simple-prod-collector-5b788bdcbb-5szz7 Binding Scheduled Successfully assigned kuttl-test-honest-termite/simple-prod-collector-5b788bdcbb-5szz7 to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:41 +0000 UTC Normal Pod simple-prod-collector-5b788bdcbb-drg4t Binding Scheduled Successfully assigned kuttl-test-honest-termite/simple-prod-collector-5b788bdcbb-drg4t to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:41 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-5b788bdcbb SuccessfulCreate Created pod: simple-prod-collector-5b788bdcbb-drg4t replicaset-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:41 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-5b788bdcbb SuccessfulCreate Created pod: simple-prod-collector-5b788bdcbb-5szz7 replicaset-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:41 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-5b788bdcbb to 2 deployment-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:41 +0000 UTC Normal Pod simple-prod-query-96c447c8d-8sfvn Binding Scheduled Successfully assigned kuttl-test-honest-termite/simple-prod-query-96c447c8d-8sfvn to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:41 +0000 UTC Normal Pod simple-prod-query-96c447c8d-st8v2 Binding Scheduled Successfully assigned kuttl-test-honest-termite/simple-prod-query-96c447c8d-st8v2 to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:41 +0000 UTC Normal ReplicaSet.apps simple-prod-query-96c447c8d SuccessfulCreate Created pod: simple-prod-query-96c447c8d-8sfvn replicaset-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:41 +0000 UTC Normal ReplicaSet.apps simple-prod-query-96c447c8d SuccessfulCreate Created pod: simple-prod-query-96c447c8d-st8v2 replicaset-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:41 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-96c447c8d to 2 deployment-controller logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:42 +0000 UTC Warning Pod simple-prod-collector-5b788bdcbb-5szz7 FailedMount MountVolume.SetUp failed for volume "simple-prod-trusted-ca" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:42 +0000 UTC Warning Pod simple-prod-collector-5b788bdcbb-5szz7 FailedMount MountVolume.SetUp failed for volume "simple-prod-sampling-configuration-volume" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:42 +0000 UTC Warning Pod simple-prod-collector-5b788bdcbb-5szz7 FailedMount MountVolume.SetUp failed for volume "simple-prod-collector-tls-config-volume" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:42 +0000 UTC Warning Pod simple-prod-collector-5b788bdcbb-5szz7 FailedMount MountVolume.SetUp failed for volume "certs" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:42 +0000 UTC Normal Pod simple-prod-collector-5b788bdcbb-drg4t AddedInterface Add eth0 [10.131.0.64/23] from ovn-kubernetes multus logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:42 +0000 UTC Normal Pod simple-prod-collector-5b788bdcbb-drg4t.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:42 +0000 UTC Normal Pod simple-prod-collector-5b788bdcbb-drg4t.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:42 +0000 UTC Normal Pod simple-prod-collector-5b788bdcbb-drg4t.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:42 +0000 UTC Normal Pod simple-prod-query-96c447c8d-8sfvn AddedInterface Add eth0 [10.131.0.65/23] from ovn-kubernetes multus logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:42 +0000 UTC Normal Pod simple-prod-query-96c447c8d-8sfvn.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:42 +0000 UTC Normal Pod simple-prod-query-96c447c8d-8sfvn.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:42 +0000 UTC Normal Pod simple-prod-query-96c447c8d-8sfvn.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:42 +0000 UTC Normal Pod simple-prod-query-96c447c8d-8sfvn.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:42 +0000 UTC Normal Pod simple-prod-query-96c447c8d-8sfvn.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:42 +0000 UTC Normal Pod simple-prod-query-96c447c8d-8sfvn.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:42 +0000 UTC Normal Pod simple-prod-query-96c447c8d-8sfvn.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:42 +0000 UTC Normal Pod simple-prod-query-96c447c8d-8sfvn.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:42 +0000 UTC Normal Pod simple-prod-query-96c447c8d-8sfvn.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:42 +0000 UTC Normal Pod simple-prod-query-96c447c8d-st8v2 AddedInterface Add eth0 [10.129.2.64/23] from ovn-kubernetes multus logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:42 +0000 UTC Normal Pod simple-prod-query-96c447c8d-st8v2.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:42 +0000 UTC Normal Pod simple-prod-query-96c447c8d-st8v2.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:42 +0000 UTC Normal Pod simple-prod-query-96c447c8d-st8v2.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:42 +0000 UTC Normal Pod simple-prod-query-96c447c8d-st8v2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:42 +0000 UTC Normal Pod simple-prod-query-96c447c8d-st8v2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:42 +0000 UTC Normal Pod simple-prod-query-96c447c8d-st8v2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:42 +0000 UTC Normal Pod simple-prod-query-96c447c8d-st8v2.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:43 +0000 UTC Normal Pod simple-prod-collector-5b788bdcbb-5szz7 AddedInterface Add eth0 [10.128.2.32/23] from ovn-kubernetes multus logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:43 +0000 UTC Normal Pod simple-prod-collector-5b788bdcbb-5szz7.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:43 +0000 UTC Normal Pod simple-prod-query-96c447c8d-st8v2.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:43 +0000 UTC Normal Pod simple-prod-query-96c447c8d-st8v2.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:44 +0000 UTC Normal Pod simple-prod-collector-5b788bdcbb-5szz7.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 18:09:44 | es-increasing-replicas | 2024-09-30 18:09:44 +0000 UTC Normal Pod simple-prod-collector-5b788bdcbb-5szz7.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 18:09:44 | es-increasing-replicas | Deleting namespace: kuttl-test-honest-termite === CONT kuttl/harness/es-index-cleaner-autoprov logger.go:42: 18:10:19 | es-index-cleaner-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 18:10:19 | es-index-cleaner-autoprov | Creating namespace: kuttl-test-famous-orca logger.go:42: 18:10:19 | es-index-cleaner-autoprov/1-install | starting test step 1-install logger.go:42: 18:10:19 | es-index-cleaner-autoprov/1-install | Jaeger:kuttl-test-famous-orca/test-es-index-cleaner-with-prefix created logger.go:42: 18:10:55 | es-index-cleaner-autoprov/1-install | test step completed 1-install logger.go:42: 18:10:55 | es-index-cleaner-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 18:10:55 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE test-es-index-cleaner-with-prefix /dev/null] logger.go:42: 18:10:57 | es-index-cleaner-autoprov/2-report-spans | Warning: resource jaegers/test-es-index-cleaner-with-prefix is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 18:11:03 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c DAYS=5 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 18:11:03 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 18:11:04 | es-index-cleaner-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 18:21:04 | es-index-cleaner-autoprov/2-report-spans | test step failed 2-report-spans case.go:364: failed in step 2-report-spans case.go:366: --- Job:kuttl-test-famous-orca/00-report-span +++ Job:kuttl-test-famous-orca/00-report-span @@ -1,8 +1,162 @@ apiVersion: batch/v1 kind: Job metadata: + annotations: + kubectl.kubernetes.io/last-applied-configuration: | + {"apiVersion":"batch/v1","kind":"Job","metadata":{"annotations":{},"name":"00-report-span","namespace":"kuttl-test-famous-orca"},"spec":{"backoffLimit":15,"template":{"spec":{"containers":[{"command":["./reporter","--days","5","--verbose"],"env":[{"name":"JAEGER_SERVICE_NAME","value":"smoke-test-service"},{"name":"OPERATION_NAME","value":"smoke-test-operation"},{"name":"JAEGER_ENDPOINT","value":"http://test-es-index-cleaner-with-prefix-collector-headless:14268/api/traces"},{"name":"JAEGER_QUERY","value":"https://test-es-index-cleaner-with-prefix-query/api/traces"},{"name":"SECRET_PATH","value":"/var/run/secrets/api-token/token"}],"image":"quay.io/rhn_support_ikanse/jaeger-asserts:latest","name":"asserts-container","volumeMounts":[{"mountPath":"/var/run/secrets/api-token","name":"token-api-volume"}]}],"restartPolicy":"OnFailure","volumes":[{"name":"token-api-volume","secret":{"secretName":"e2e-test"}}]}}}} + labels: + batch.kubernetes.io/controller-uid: 012cde23-d26b-4583-b77b-d4ed08b251de + batch.kubernetes.io/job-name: 00-report-span + controller-uid: 012cde23-d26b-4583-b77b-d4ed08b251de + job-name: 00-report-span + managedFields: + - apiVersion: batch/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + .: {} + f:kubectl.kubernetes.io/last-applied-configuration: {} + f:spec: + f:backoffLimit: {} + f:completionMode: {} + f:completions: {} + f:manualSelector: {} + f:parallelism: {} + f:podReplacementPolicy: {} + f:suspend: {} + f:template: + f:spec: + f:containers: + k:{"name":"asserts-container"}: + .: {} + f:command: {} + f:env: + .: {} + k:{"name":"JAEGER_ENDPOINT"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"JAEGER_QUERY"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"JAEGER_SERVICE_NAME"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"OPERATION_NAME"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"SECRET_PATH"}: + .: {} + f:name: {} + f:value: {} + f:image: {} + f:imagePullPolicy: {} + f:name: {} + f:resources: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/var/run/secrets/api-token"}: + .: {} + f:mountPath: {} + f:name: {} + f:dnsPolicy: {} + f:restartPolicy: {} + f:schedulerName: {} + f:securityContext: {} + f:terminationGracePeriodSeconds: {} + f:volumes: + .: {} + k:{"name":"token-api-volume"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + manager: kubectl-client-side-apply + operation: Update + time: "2024-09-30T18:11:04Z" + - apiVersion: batch/v1 + fieldsType: FieldsV1 + fieldsV1: + f:status: + f:active: {} + f:ready: {} + f:startTime: {} + f:terminating: {} + f:uncountedTerminatedPods: {} + manager: kube-controller-manager + operation: Update + subresource: status + time: "2024-09-30T18:11:06Z" name: 00-report-span namespace: kuttl-test-famous-orca +spec: + backoffLimit: 15 + completionMode: NonIndexed + completions: 1 + manualSelector: false + parallelism: 1 + podReplacementPolicy: TerminatingOrFailed + selector: + matchLabels: + batch.kubernetes.io/controller-uid: 012cde23-d26b-4583-b77b-d4ed08b251de + suspend: false + template: + metadata: + creationTimestamp: null + labels: + batch.kubernetes.io/controller-uid: 012cde23-d26b-4583-b77b-d4ed08b251de + batch.kubernetes.io/job-name: 00-report-span + controller-uid: 012cde23-d26b-4583-b77b-d4ed08b251de + job-name: 00-report-span + spec: + containers: + - command: + - ./reporter + - --days + - "5" + - --verbose + env: + - name: JAEGER_SERVICE_NAME + value: smoke-test-service + - name: OPERATION_NAME + value: smoke-test-operation + - name: JAEGER_ENDPOINT + value: http://test-es-index-cleaner-with-prefix-collector-headless:14268/api/traces + - name: JAEGER_QUERY + value: https://test-es-index-cleaner-with-prefix-query/api/traces + - name: SECRET_PATH + value: /var/run/secrets/api-token/token + image: quay.io/rhn_support_ikanse/jaeger-asserts:latest + imagePullPolicy: Always + name: asserts-container + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /var/run/secrets/api-token + name: token-api-volume + dnsPolicy: ClusterFirst + restartPolicy: OnFailure + schedulerName: default-scheduler + securityContext: {} + terminationGracePeriodSeconds: 30 + volumes: + - name: token-api-volume + secret: + defaultMode: 420 + secretName: e2e-test status: - succeeded: 1 + active: 1 + ready: 1 + startTime: "2024-09-30T18:11:04Z" + terminating: 0 + uncountedTerminatedPods: {} case.go:366: resource Job:kuttl-test-famous-orca/00-report-span: .status.succeeded: key is missing from map logger.go:42: 18:21:04 | es-index-cleaner-autoprov | es-index-cleaner-autoprov events from ns kuttl-test-famous-orca: logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:25 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestfamousorcatestesindexcleane-1-6b776d8d8 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestfamousorcatestesindexcleane-1-6jtlzb replicaset-controller logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfamousorcatestesindexcleane-1-6jtlzb Binding Scheduled Successfully assigned kuttl-test-famous-orca/elasticsearch-cdm-kuttltestfamousorcatestesindexcleane-1-6jtlzb to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:25 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestfamousorcatestesindexcleane-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestfamousorcatestesindexcleane-1-6b776d8d8 to 1 deployment-controller logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfamousorcatestesindexcleane-1-6jtlzb AddedInterface Add eth0 [10.128.2.33/23] from ovn-kubernetes multus logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfamousorcatestesindexcleane-1-6jtlzb.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:89629964e44058bafc90393a235b08c7c974f05513aecbfe7229134da732f2b5" already present on machine kubelet logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfamousorcatestesindexcleane-1-6jtlzb.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfamousorcatestesindexcleane-1-6jtlzb.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfamousorcatestesindexcleane-1-6jtlzb.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:a8f53dec42a46c5bf8ac7f2888848c01e70f405d27211e2a269730c888929faf" already present on machine kubelet logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfamousorcatestesindexcleane-1-6jtlzb.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfamousorcatestesindexcleane-1-6jtlzb.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:36 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestfamousorcatestesindexcleane-1-6jtlzb.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:41 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestfamousorcatestesindexcleane-1-6jtlzb.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:52 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-78f8b7fc9d-b2qbd Binding Scheduled Successfully assigned kuttl-test-famous-orca/test-es-index-cleaner-with-prefix-collector-78f8b7fc9d-b2qbd to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:52 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-collector-78f8b7fc9d SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-collector-78f8b7fc9d-b2qbd replicaset-controller logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:52 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-collector ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-collector-78f8b7fc9d to 1 deployment-controller logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:52 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-c7dfc77d5-pgqzf Binding Scheduled Successfully assigned kuttl-test-famous-orca/test-es-index-cleaner-with-prefix-query-c7dfc77d5-pgqzf to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:52 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-c7dfc77d5 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-query-c7dfc77d5-pgqzf replicaset-controller logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:52 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-query-c7dfc77d5 to 1 deployment-controller logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-78f8b7fc9d-b2qbd AddedInterface Add eth0 [10.129.2.65/23] from ovn-kubernetes multus logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-78f8b7fc9d-b2qbd.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-78f8b7fc9d-b2qbd.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-78f8b7fc9d-b2qbd.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-c7dfc77d5-pgqzf AddedInterface Add eth0 [10.131.0.66/23] from ovn-kubernetes multus logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-c7dfc77d5-pgqzf.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-c7dfc77d5-pgqzf.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-c7dfc77d5-pgqzf.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-c7dfc77d5-pgqzf.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-c7dfc77d5-pgqzf.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-c7dfc77d5-pgqzf.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-c7dfc77d5-pgqzf.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-c7dfc77d5-pgqzf.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:10:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-c7dfc77d5-pgqzf.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:11:04 +0000 UTC Normal Pod 00-report-span-g5mqz Binding Scheduled Successfully assigned kuttl-test-famous-orca/00-report-span-g5mqz to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:11:04 +0000 UTC Normal Pod 00-report-span-g5mqz AddedInterface Add eth0 [10.129.2.66/23] from ovn-kubernetes multus logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:11:04 +0000 UTC Normal Pod 00-report-span-g5mqz.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:11:04 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-g5mqz job-controller logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:11:05 +0000 UTC Normal Pod 00-report-span-g5mqz.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 228ms (228ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:11:05 +0000 UTC Normal Pod 00-report-span-g5mqz.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:11:05 +0000 UTC Normal Pod 00-report-span-g5mqz.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:11:10 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:11:10 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:11:10 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:11:25 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:11:25 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-78f8b7fc9d-b2qbd horizontal-pod-autoscaler logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:11:25 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:11:40 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-78f8b7fc9d-b2qbd horizontal-pod-autoscaler logger.go:42: 18:21:04 | es-index-cleaner-autoprov | 2024-09-30 18:11:40 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-78f8b7fc9d-b2qbd horizontal-pod-autoscaler logger.go:42: 18:21:04 | es-index-cleaner-autoprov | Deleting namespace: kuttl-test-famous-orca === CONT kuttl/harness/es-from-aio-to-production logger.go:42: 18:21:11 | es-from-aio-to-production | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 18:21:11 | es-from-aio-to-production | Creating namespace: kuttl-test-teaching-poodle logger.go:42: 18:21:11 | es-from-aio-to-production/0-install | starting test step 0-install logger.go:42: 18:21:12 | es-from-aio-to-production/0-install | Jaeger:kuttl-test-teaching-poodle/my-jaeger created logger.go:42: 18:21:18 | es-from-aio-to-production/0-install | test step completed 0-install logger.go:42: 18:21:18 | es-from-aio-to-production/1-smoke-test | starting test step 1-smoke-test logger.go:42: 18:21:18 | es-from-aio-to-production/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 18:21:19 | es-from-aio-to-production/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 18:21:26 | es-from-aio-to-production/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 18:21:26 | es-from-aio-to-production/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 18:21:26 | es-from-aio-to-production/1-smoke-test | job.batch/report-span created logger.go:42: 18:21:26 | es-from-aio-to-production/1-smoke-test | job.batch/check-span created logger.go:42: 18:21:39 | es-from-aio-to-production/1-smoke-test | test step completed 1-smoke-test logger.go:42: 18:21:39 | es-from-aio-to-production/3-install | starting test step 3-install logger.go:42: 18:21:39 | es-from-aio-to-production/3-install | Jaeger:kuttl-test-teaching-poodle/my-jaeger updated logger.go:42: 18:22:12 | es-from-aio-to-production/3-install | test step completed 3-install logger.go:42: 18:22:12 | es-from-aio-to-production/4-smoke-test | starting test step 4-smoke-test logger.go:42: 18:22:12 | es-from-aio-to-production/4-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 18:22:20 | es-from-aio-to-production/4-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 18:22:20 | es-from-aio-to-production/4-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 18:22:20 | es-from-aio-to-production/4-smoke-test | job.batch/report-span unchanged logger.go:42: 18:22:20 | es-from-aio-to-production/4-smoke-test | job.batch/check-span unchanged logger.go:42: 18:22:21 | es-from-aio-to-production/4-smoke-test | test step completed 4-smoke-test logger.go:42: 18:22:21 | es-from-aio-to-production | es-from-aio-to-production events from ns kuttl-test-teaching-poodle: logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:15 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-6ddcf78897 to 1 deployment-controller logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:16 +0000 UTC Normal Pod my-jaeger-6ddcf78897-7ctmr Binding Scheduled Successfully assigned kuttl-test-teaching-poodle/my-jaeger-6ddcf78897-7ctmr to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:16 +0000 UTC Normal Pod my-jaeger-6ddcf78897-7ctmr AddedInterface Add eth0 [10.131.0.68/23] from ovn-kubernetes multus logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:16 +0000 UTC Normal Pod my-jaeger-6ddcf78897-7ctmr.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:16 +0000 UTC Normal Pod my-jaeger-6ddcf78897-7ctmr.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:16 +0000 UTC Normal Pod my-jaeger-6ddcf78897-7ctmr.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:16 +0000 UTC Normal Pod my-jaeger-6ddcf78897-7ctmr.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:16 +0000 UTC Normal Pod my-jaeger-6ddcf78897-7ctmr.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:16 +0000 UTC Normal Pod my-jaeger-6ddcf78897-7ctmr.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:16 +0000 UTC Normal ReplicaSet.apps my-jaeger-6ddcf78897 SuccessfulCreate Created pod: my-jaeger-6ddcf78897-7ctmr replicaset-controller logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:23 +0000 UTC Normal Pod my-jaeger-6ddcf78897-7ctmr.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:23 +0000 UTC Normal Pod my-jaeger-6ddcf78897-7ctmr.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:23 +0000 UTC Normal ReplicaSet.apps my-jaeger-6ddcf78897 SuccessfulDelete Deleted pod: my-jaeger-6ddcf78897-7ctmr replicaset-controller logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:23 +0000 UTC Normal Pod my-jaeger-858bb885c5-rsr2h Binding Scheduled Successfully assigned kuttl-test-teaching-poodle/my-jaeger-858bb885c5-rsr2h to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:23 +0000 UTC Normal ReplicaSet.apps my-jaeger-858bb885c5 SuccessfulCreate Created pod: my-jaeger-858bb885c5-rsr2h replicaset-controller logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:23 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-6ddcf78897 to 0 from 1 deployment-controller logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:23 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-858bb885c5 to 1 deployment-controller logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:24 +0000 UTC Normal Pod my-jaeger-858bb885c5-rsr2h AddedInterface Add eth0 [10.131.0.69/23] from ovn-kubernetes multus logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:24 +0000 UTC Normal Pod my-jaeger-858bb885c5-rsr2h.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:24 +0000 UTC Normal Pod my-jaeger-858bb885c5-rsr2h.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:24 +0000 UTC Normal Pod my-jaeger-858bb885c5-rsr2h.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:24 +0000 UTC Normal Pod my-jaeger-858bb885c5-rsr2h.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:24 +0000 UTC Normal Pod my-jaeger-858bb885c5-rsr2h.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:24 +0000 UTC Normal Pod my-jaeger-858bb885c5-rsr2h.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:26 +0000 UTC Normal Pod check-span-2h4nf Binding Scheduled Successfully assigned kuttl-test-teaching-poodle/check-span-2h4nf to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:26 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-2h4nf job-controller logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:26 +0000 UTC Normal Pod report-span-s9ll9 Binding Scheduled Successfully assigned kuttl-test-teaching-poodle/report-span-s9ll9 to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:26 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-s9ll9 job-controller logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:27 +0000 UTC Normal Pod check-span-2h4nf AddedInterface Add eth0 [10.129.2.67/23] from ovn-kubernetes multus logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:27 +0000 UTC Normal Pod check-span-2h4nf.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:27 +0000 UTC Normal Pod check-span-2h4nf.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 229ms (229ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:27 +0000 UTC Normal Pod check-span-2h4nf.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:27 +0000 UTC Normal Pod check-span-2h4nf.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:27 +0000 UTC Normal Pod report-span-s9ll9 AddedInterface Add eth0 [10.128.2.34/23] from ovn-kubernetes multus logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:27 +0000 UTC Normal Pod report-span-s9ll9.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:28 +0000 UTC Normal Pod report-span-s9ll9.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 808ms (808ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:28 +0000 UTC Normal Pod report-span-s9ll9.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:28 +0000 UTC Normal Pod report-span-s9ll9.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:38 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:42 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestteachingpoodlemyjaeger-1-dd755b876 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestteachingpoodlemyjaeger-1-dd755bf2mql replicaset-controller logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingpoodlemyjaeger-1-dd755bf2mql Binding Scheduled Successfully assigned kuttl-test-teaching-poodle/elasticsearch-cdm-kuttltestteachingpoodlemyjaeger-1-dd755bf2mql to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:42 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestteachingpoodlemyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestteachingpoodlemyjaeger-1-dd755b876 to 1 deployment-controller logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingpoodlemyjaeger-1-dd755bf2mql AddedInterface Add eth0 [10.129.2.68/23] from ovn-kubernetes multus logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingpoodlemyjaeger-1-dd755bf2mql.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:89629964e44058bafc90393a235b08c7c974f05513aecbfe7229134da732f2b5" already present on machine kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingpoodlemyjaeger-1-dd755bf2mql.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingpoodlemyjaeger-1-dd755bf2mql.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingpoodlemyjaeger-1-dd755bf2mql.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:a8f53dec42a46c5bf8ac7f2888848c01e70f405d27211e2a269730c888929faf" already present on machine kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingpoodlemyjaeger-1-dd755bf2mql.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestteachingpoodlemyjaeger-1-dd755bf2mql.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:53 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestteachingpoodlemyjaeger-1-dd755bf2mql.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:21:58 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestteachingpoodlemyjaeger-1-dd755bf2mql.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:22:00 +0000 UTC Normal Job.batch report-span Completed Job completed job-controller logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:22:09 +0000 UTC Normal Pod my-jaeger-858bb885c5-rsr2h.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:22:09 +0000 UTC Normal Pod my-jaeger-858bb885c5-rsr2h.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:22:09 +0000 UTC Normal Pod my-jaeger-collector-556f7bdb54-gngpr Binding Scheduled Successfully assigned kuttl-test-teaching-poodle/my-jaeger-collector-556f7bdb54-gngpr to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:22:09 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-556f7bdb54 SuccessfulCreate Created pod: my-jaeger-collector-556f7bdb54-gngpr replicaset-controller logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:22:09 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-556f7bdb54 to 1 deployment-controller logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:22:09 +0000 UTC Normal Pod my-jaeger-query-85d65cc4fb-rfxzq Binding Scheduled Successfully assigned kuttl-test-teaching-poodle/my-jaeger-query-85d65cc4fb-rfxzq to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:22:09 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-85d65cc4fb SuccessfulCreate Created pod: my-jaeger-query-85d65cc4fb-rfxzq replicaset-controller logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:22:09 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-85d65cc4fb to 1 deployment-controller logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:22:10 +0000 UTC Normal Pod my-jaeger-collector-556f7bdb54-gngpr AddedInterface Add eth0 [10.128.2.35/23] from ovn-kubernetes multus logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:22:10 +0000 UTC Normal Pod my-jaeger-collector-556f7bdb54-gngpr.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:22:10 +0000 UTC Normal Pod my-jaeger-collector-556f7bdb54-gngpr.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:22:10 +0000 UTC Normal Pod my-jaeger-collector-556f7bdb54-gngpr.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:22:10 +0000 UTC Normal Pod my-jaeger-query-85d65cc4fb-rfxzq AddedInterface Add eth0 [10.131.0.70/23] from ovn-kubernetes multus logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:22:10 +0000 UTC Normal Pod my-jaeger-query-85d65cc4fb-rfxzq.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:22:10 +0000 UTC Normal Pod my-jaeger-query-85d65cc4fb-rfxzq.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:22:10 +0000 UTC Normal Pod my-jaeger-query-85d65cc4fb-rfxzq.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:22:10 +0000 UTC Normal Pod my-jaeger-query-85d65cc4fb-rfxzq.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:22:10 +0000 UTC Normal Pod my-jaeger-query-85d65cc4fb-rfxzq.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:22:10 +0000 UTC Normal Pod my-jaeger-query-85d65cc4fb-rfxzq.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:22:10 +0000 UTC Normal Pod my-jaeger-query-85d65cc4fb-rfxzq.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:22:10 +0000 UTC Normal Pod my-jaeger-query-85d65cc4fb-rfxzq.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | 2024-09-30 18:22:10 +0000 UTC Normal Pod my-jaeger-query-85d65cc4fb-rfxzq.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 18:22:21 | es-from-aio-to-production | Deleting namespace: kuttl-test-teaching-poodle === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- FAIL: kuttl (1182.16s) --- FAIL: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.93s) --- PASS: kuttl/harness/es-multiinstance (98.22s) --- PASS: kuttl/harness/es-simple-prod (5.96s) --- PASS: kuttl/harness/es-rollover-autoprov (232.22s) --- PASS: kuttl/harness/es-increasing-replicas (111.56s) --- FAIL: kuttl/harness/es-index-cleaner-autoprov (652.50s) --- PASS: kuttl/harness/es-from-aio-to-production (75.60s) FAIL + exit_code=1 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name elasticsearch --report --output /logs/artifacts/elasticsearch.xml ./artifacts/kuttl-report.xml time="2024-09-30T18:22:28Z" level=debug msg="Setting a new name for the test suites" time="2024-09-30T18:22:28Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-09-30T18:22:28Z" level=debug msg="normalizing test case names" time="2024-09-30T18:22:28Z" level=debug msg="elasticsearch/artifacts -> elasticsearch_artifacts" time="2024-09-30T18:22:28Z" level=debug msg="elasticsearch/es-multiinstance -> elasticsearch_es_multiinstance" time="2024-09-30T18:22:28Z" level=debug msg="elasticsearch/es-simple-prod -> elasticsearch_es_simple_prod" time="2024-09-30T18:22:28Z" level=debug msg="elasticsearch/es-rollover-autoprov -> elasticsearch_es_rollover_autoprov" time="2024-09-30T18:22:28Z" level=debug msg="elasticsearch/es-increasing-replicas -> elasticsearch_es_increasing_replicas" time="2024-09-30T18:22:28Z" level=debug msg="elasticsearch/es-index-cleaner-autoprov -> elasticsearch_es_index_cleaner_autoprov" time="2024-09-30T18:22:28Z" level=debug msg="elasticsearch/es-from-aio-to-production -> elasticsearch_es_from_aio_to_production" +-----------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------+--------+ | elasticsearch_artifacts | passed | | elasticsearch_es_multiinstance | passed | | elasticsearch_es_simple_prod | passed | | elasticsearch_es_rollover_autoprov | passed | | elasticsearch_es_increasing_replicas | passed | | elasticsearch_es_index_cleaner_autoprov | failed | | elasticsearch_es_from_aio_to_production | passed | +-----------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh examples false true + '[' 3 -ne 3 ']' + test_suite_name=examples + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/examples.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-examples make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ VERTX_IMG=jaegertracing/vertx-create-span:operator-e2e-tests \ ./tests/e2e/examples/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.17.0-0.nightly-2024-09-29-173421 True False 45m Cluster version is 4.17.0-0.nightly-2024-09-29-173421' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.17.0-0.nightly-2024-09-29-173421 True False 45m Cluster version is 4.17.0-0.nightly-2024-09-29-173421' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/examples/render.sh ++ export SUITE_DIR=./tests/e2e/examples ++ SUITE_DIR=./tests/e2e/examples ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/examples ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test examples-agent-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-agent-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-with-priority-class\e[0m' Rendering files for test examples-agent-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + mkdir -p examples-agent-with-priority-class + cd examples-agent-with-priority-class + example_name=agent-with-priority-class + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + '[' true '!=' true ']' + render_install_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + install_file=./02-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/agent-with-priority-class.yaml -o ./02-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./02-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./02-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./02-install.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + JAEGER_NAME=agent-as-daemonset + local jaeger_strategy ++ get_jaeger_strategy ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./02-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./02-install.yaml ++ strategy=DaemonSet ++ '[' DaemonSet = null ']' ++ echo DaemonSet ++ return 0 + jaeger_strategy=DaemonSet + '[' DaemonSet = DaemonSet ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./02-assert.yaml + render_smoke_test_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + jaeger_name=agent-as-daemonset + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test agent-as-daemonset true 02 + '[' 3 -ne 3 ']' + jaeger=agent-as-daemonset + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + export JAEGER_NAME=agent-as-daemonset + JAEGER_NAME=agent-as-daemonset + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-all-in-one-with-options + '[' 1 -ne 1 ']' + test_name=examples-all-in-one-with-options + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-all-in-one-with-options' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-all-in-one-with-options\e[0m' Rendering files for test examples-all-in-one-with-options + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-with-priority-class + '[' examples-agent-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-all-in-one-with-options + cd examples-all-in-one-with-options + example_name=all-in-one-with-options + render_install_example all-in-one-with-options 00 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/all-in-one-with-options.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + JAEGER_NAME=my-jaeger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.metadata.name="my-jaeger"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i 'del(.spec.allInOne.image)' ./00-install.yaml + render_smoke_test_example all-in-one-with-options 01 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + jaeger_name=my-jaeger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + sed -i s~my-jaeger-query:443~my-jaeger-query:443/jaeger~gi ./01-smoke-test.yaml + start_test examples-business-application-injected-sidecar + '[' 1 -ne 1 ']' + test_name=examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-business-application-injected-sidecar' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-business-application-injected-sidecar\e[0m' Rendering files for test examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-all-in-one-with-options + '[' examples-all-in-one-with-options '!=' _build ']' + cd .. + mkdir -p examples-business-application-injected-sidecar + cd examples-business-application-injected-sidecar + example_name=simplest + cp /tmp/jaeger-tests/examples/business-application-injected-sidecar.yaml ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].image=strenv(VERTX_IMG)' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.path="/"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.port=8080' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.initialDelaySeconds=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.failureThreshold=3' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.periodSeconds=10' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.successThreshold=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.timeoutSeconds=1' ./00-install.yaml + render_install_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example simplest 02 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 02 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-collector-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-collector-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-collector-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-collector-with-priority-class\e[0m' Rendering files for test examples-collector-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-business-application-injected-sidecar + '[' examples-business-application-injected-sidecar '!=' _build ']' + cd .. + mkdir -p examples-collector-with-priority-class + cd examples-collector-with-priority-class + example_name=collector-with-priority-class + render_install_example collector-with-priority-class 00 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/collector-with-priority-class.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + JAEGER_NAME=collector-with-high-priority + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example collector-with-priority-class 01 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + jaeger_name=collector-with-high-priority + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test collector-with-high-priority true 01 + '[' 3 -ne 3 ']' + jaeger=collector-with-high-priority + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + export JAEGER_NAME=collector-with-high-priority + JAEGER_NAME=collector-with-high-priority + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-service-types + '[' 1 -ne 1 ']' + test_name=examples-service-types + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-service-types' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-service-types\e[0m' Rendering files for test examples-service-types + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-collector-with-priority-class + '[' examples-collector-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-service-types + cd examples-service-types + example_name=service-types + render_install_example service-types 00 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/service-types.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + JAEGER_NAME=service-types + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example service-types 01 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/service-types.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/service-types.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/service-types.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/service-types.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + jaeger_name=service-types + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test service-types true 01 + '[' 3 -ne 3 ']' + jaeger=service-types + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + export JAEGER_NAME=service-types + JAEGER_NAME=service-types + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod + '[' 1 -ne 1 ']' + test_name=examples-simple-prod + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod\e[0m' Rendering files for test examples-simple-prod + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-service-types + '[' examples-service-types '!=' _build ']' + cd .. + mkdir -p examples-simple-prod + cd examples-simple-prod + example_name=simple-prod + render_install_example simple-prod 01 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod 02 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod-with-volumes + '[' 1 -ne 1 ']' + test_name=examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod-with-volumes' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod-with-volumes\e[0m' Rendering files for test examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod + '[' examples-simple-prod '!=' _build ']' + cd .. + mkdir -p examples-simple-prod-with-volumes + cd examples-simple-prod-with-volumes + example_name=simple-prod-with-volumes + render_install_example simple-prod-with-volumes 01 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod-with-volumes 02 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + /tmp/jaeger-tests/bin/gomplate -f ./03-check-volume.yaml.template -o 03-check-volume.yaml + start_test examples-simplest + '[' 1 -ne 1 ']' + test_name=examples-simplest + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simplest' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simplest\e[0m' Rendering files for test examples-simplest + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod-with-volumes + '[' examples-simple-prod-with-volumes '!=' _build ']' + cd .. + mkdir -p examples-simplest + cd examples-simplest + example_name=simplest + render_install_example simplest 00 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 01 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger + '[' 1 -ne 1 ']' + test_name=examples-with-badger + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger\e[0m' Rendering files for test examples-with-badger + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simplest + '[' examples-simplest '!=' _build ']' + cd .. + mkdir -p examples-with-badger + cd examples-with-badger + example_name=with-badger + render_install_example with-badger 00 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + JAEGER_NAME=with-badger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger 01 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + jaeger_name=with-badger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + export JAEGER_NAME=with-badger + JAEGER_NAME=with-badger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger-and-volume + '[' 1 -ne 1 ']' + test_name=examples-with-badger-and-volume + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger-and-volume' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger-and-volume\e[0m' Rendering files for test examples-with-badger-and-volume + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger + '[' examples-with-badger '!=' _build ']' + cd .. + mkdir -p examples-with-badger-and-volume + cd examples-with-badger-and-volume + example_name=with-badger-and-volume + render_install_example with-badger-and-volume 00 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger-and-volume.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + JAEGER_NAME=with-badger-and-volume + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger-and-volume 01 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + jaeger_name=with-badger-and-volume + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger-and-volume true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger-and-volume + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + export JAEGER_NAME=with-badger-and-volume + JAEGER_NAME=with-badger-and-volume + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-cassandra + '[' 1 -ne 1 ']' + test_name=examples-with-cassandra + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-cassandra' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-cassandra\e[0m' Rendering files for test examples-with-cassandra + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger-and-volume + '[' examples-with-badger-and-volume '!=' _build ']' + cd .. + mkdir -p examples-with-cassandra + cd examples-with-cassandra + example_name=with-cassandra + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-cassandra 01 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-cassandra.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + JAEGER_NAME=with-cassandra + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-cassandra 02 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-cassandra.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-cassandra.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + jaeger_name=with-cassandra + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-cassandra true 02 + '[' 3 -ne 3 ']' + jaeger=with-cassandra + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + export JAEGER_NAME=with-cassandra + JAEGER_NAME=with-cassandra + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-sampling + '[' 1 -ne 1 ']' + test_name=examples-with-sampling + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-sampling' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-sampling\e[0m' Rendering files for test examples-with-sampling + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-cassandra + '[' examples-with-cassandra '!=' _build ']' + cd .. + mkdir -p examples-with-sampling + cd examples-with-sampling + export example_name=with-sampling + example_name=with-sampling + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-sampling 01 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-sampling.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + JAEGER_NAME=with-sampling + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-sampling 02 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-sampling.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-sampling.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + jaeger_name=with-sampling + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-sampling true 02 + '[' 3 -ne 3 ']' + jaeger=with-sampling + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + export JAEGER_NAME=with-sampling + JAEGER_NAME=with-sampling + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-agent-as-daemonset + '[' 1 -ne 1 ']' + test_name=examples-agent-as-daemonset + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-as-daemonset' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-as-daemonset\e[0m' Rendering files for test examples-agent-as-daemonset + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-sampling + '[' examples-with-sampling '!=' _build ']' + cd .. + mkdir -p examples-agent-as-daemonset + cd examples-agent-as-daemonset + '[' true = true ']' + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/agent-as-daemonset.yaml -o 02-install.yaml + '[' true = true ']' + start_test examples-openshift-with-htpasswd + '[' 1 -ne 1 ']' + test_name=examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-openshift-with-htpasswd' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-openshift-with-htpasswd\e[0m' Rendering files for test examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-as-daemonset + '[' examples-agent-as-daemonset '!=' _build ']' + cd .. + mkdir -p examples-openshift-with-htpasswd + cd examples-openshift-with-htpasswd + export JAEGER_NAME=with-htpasswd + JAEGER_NAME=with-htpasswd + export JAEGER_USERNAME=awesomeuser + JAEGER_USERNAME=awesomeuser + export JAEGER_PASSWORD=awesomepassword + JAEGER_PASSWORD=awesomepassword + export 'JAEGER_USER_PASSWORD_HASH=awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' + JAEGER_USER_PASSWORD_HASH='awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ echo 'awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ base64 + SECRET=YXdlc29tZXVzZXI6e1NIQX11VWRxUFZVeXFOQm1FUlUwUXhqM0tGYVpuanc9Cg== + /tmp/jaeger-tests/bin/gomplate -f ./00-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/with-htpasswd.yaml -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + export 'GET_URL_COMMAND=kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + export 'URL=https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + INSECURE=true + JAEGER_USERNAME= + JAEGER_PASSWORD= + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-unsecured.yaml + JAEGER_USERNAME=wronguser + JAEGER_PASSWORD=wrongpassword + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-check-unauthorized.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./04-check-authorized.yaml + skip_test examples-agent-as-daemonset 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-agent-as-daemonset + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-openshift-with-htpasswd + '[' examples-openshift-with-htpasswd '!=' _build ']' + cd .. + rm -rf examples-agent-as-daemonset + warning 'examples-agent-as-daemonset: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-agent-as-daemonset: This test is flaky in Prow CI\e[0m' WAR: examples-agent-as-daemonset: This test is flaky in Prow CI + skip_test examples-with-badger-and-volume 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-with-badger-and-volume + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + rm -rf examples-with-badger-and-volume + warning 'examples-with-badger-and-volume: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-with-badger-and-volume: This test is flaky in Prow CI\e[0m' WAR: examples-with-badger-and-volume: This test is flaky in Prow CI + skip_test examples-collector-with-priority-class 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-collector-with-priority-class + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + rm -rf examples-collector-with-priority-class + warning 'examples-collector-with-priority-class: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-collector-with-priority-class: This test is flaky in Prow CI\e[0m' WAR: examples-collector-with-priority-class: This test is flaky in Prow CI make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running examples E2E tests' Running examples E2E tests + cd tests/e2e/examples/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2856073267 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 12 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/examples-agent-with-priority-class === PAUSE kuttl/harness/examples-agent-with-priority-class === RUN kuttl/harness/examples-all-in-one-with-options === PAUSE kuttl/harness/examples-all-in-one-with-options === RUN kuttl/harness/examples-business-application-injected-sidecar === PAUSE kuttl/harness/examples-business-application-injected-sidecar === RUN kuttl/harness/examples-openshift-with-htpasswd === PAUSE kuttl/harness/examples-openshift-with-htpasswd === RUN kuttl/harness/examples-service-types === PAUSE kuttl/harness/examples-service-types === RUN kuttl/harness/examples-simple-prod === PAUSE kuttl/harness/examples-simple-prod === RUN kuttl/harness/examples-simple-prod-with-volumes === PAUSE kuttl/harness/examples-simple-prod-with-volumes === RUN kuttl/harness/examples-simplest === PAUSE kuttl/harness/examples-simplest === RUN kuttl/harness/examples-with-badger === PAUSE kuttl/harness/examples-with-badger === RUN kuttl/harness/examples-with-cassandra === PAUSE kuttl/harness/examples-with-cassandra === RUN kuttl/harness/examples-with-sampling === PAUSE kuttl/harness/examples-with-sampling === CONT kuttl/harness/artifacts logger.go:42: 18:23:01 | artifacts | Creating namespace: kuttl-test-selected-starling logger.go:42: 18:23:01 | artifacts | artifacts events from ns kuttl-test-selected-starling: logger.go:42: 18:23:01 | artifacts | Deleting namespace: kuttl-test-selected-starling === CONT kuttl/harness/examples-simple-prod logger.go:42: 18:23:07 | examples-simple-prod | Creating namespace: kuttl-test-learning-sawfish logger.go:42: 18:23:07 | examples-simple-prod/1-install | starting test step 1-install logger.go:42: 18:23:07 | examples-simple-prod/1-install | Jaeger:kuttl-test-learning-sawfish/simple-prod created logger.go:42: 18:23:47 | examples-simple-prod/1-install | test step completed 1-install logger.go:42: 18:23:47 | examples-simple-prod/2-smoke-test | starting test step 2-smoke-test logger.go:42: 18:23:47 | examples-simple-prod/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 18:23:48 | examples-simple-prod/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 18:23:55 | examples-simple-prod/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 18:23:55 | examples-simple-prod/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 18:23:55 | examples-simple-prod/2-smoke-test | job.batch/report-span created logger.go:42: 18:23:56 | examples-simple-prod/2-smoke-test | job.batch/check-span created logger.go:42: 18:24:08 | examples-simple-prod/2-smoke-test | test step completed 2-smoke-test logger.go:42: 18:24:08 | examples-simple-prod | examples-simple-prod events from ns kuttl-test-learning-sawfish: logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:14 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestlearningsawfishsimpleprod-1-97b8fbc6 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestlearningsawfishsimpleprod-1-97bf5p7n replicaset-controller logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:14 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestlearningsawfishsimpleprod-1-97bf5p7n Binding Scheduled Successfully assigned kuttl-test-learning-sawfish/elasticsearch-cdm-kuttltestlearningsawfishsimpleprod-1-97bf5p7n to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:14 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestlearningsawfishsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestlearningsawfishsimpleprod-1-97b8fbc6 to 1 deployment-controller logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:15 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestlearningsawfishsimpleprod-1-97bf5p7n AddedInterface Add eth0 [10.129.2.69/23] from ovn-kubernetes multus logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:15 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestlearningsawfishsimpleprod-1-97bf5p7n.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:89629964e44058bafc90393a235b08c7c974f05513aecbfe7229134da732f2b5" already present on machine kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:15 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestlearningsawfishsimpleprod-1-97bf5p7n.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:15 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestlearningsawfishsimpleprod-1-97bf5p7n.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:15 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestlearningsawfishsimpleprod-1-97bf5p7n.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:a8f53dec42a46c5bf8ac7f2888848c01e70f405d27211e2a269730c888929faf" already present on machine kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:15 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestlearningsawfishsimpleprod-1-97bf5p7n.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:15 +0000 UTC Normal Pod simple-prod-collector-75f468dd87-t9crf Binding Scheduled Successfully assigned kuttl-test-learning-sawfish/simple-prod-collector-75f468dd87-t9crf to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:15 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-75f468dd87 SuccessfulCreate Created pod: simple-prod-collector-75f468dd87-t9crf replicaset-controller logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:15 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-75f468dd87 to 1 deployment-controller logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:15 +0000 UTC Normal Pod simple-prod-query-678d965c78-7q6sr Binding Scheduled Successfully assigned kuttl-test-learning-sawfish/simple-prod-query-678d965c78-7q6sr to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:15 +0000 UTC Normal ReplicaSet.apps simple-prod-query-678d965c78 SuccessfulCreate Created pod: simple-prod-query-678d965c78-7q6sr replicaset-controller logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:15 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-678d965c78 to 1 deployment-controller logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:16 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestlearningsawfishsimpleprod-1-97bf5p7n.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:16 +0000 UTC Warning Pod simple-prod-collector-75f468dd87-t9crf FailedMount MountVolume.SetUp failed for volume "simple-prod-collector-tls-config-volume" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:16 +0000 UTC Normal Pod simple-prod-query-678d965c78-7q6sr AddedInterface Add eth0 [10.131.0.72/23] from ovn-kubernetes multus logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:16 +0000 UTC Normal Pod simple-prod-query-678d965c78-7q6sr.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:17 +0000 UTC Normal Pod simple-prod-collector-75f468dd87-t9crf AddedInterface Add eth0 [10.131.0.71/23] from ovn-kubernetes multus logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:17 +0000 UTC Normal Pod simple-prod-collector-75f468dd87-t9crf.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:17 +0000 UTC Normal Pod simple-prod-collector-75f468dd87-t9crf.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:17 +0000 UTC Normal Pod simple-prod-collector-75f468dd87-t9crf.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:17 +0000 UTC Normal Pod simple-prod-query-678d965c78-7q6sr.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:17 +0000 UTC Normal Pod simple-prod-query-678d965c78-7q6sr.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:17 +0000 UTC Normal Pod simple-prod-query-678d965c78-7q6sr.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:17 +0000 UTC Normal Pod simple-prod-query-678d965c78-7q6sr.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:17 +0000 UTC Normal Pod simple-prod-query-678d965c78-7q6sr.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:17 +0000 UTC Normal Pod simple-prod-query-678d965c78-7q6sr.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:17 +0000 UTC Normal Pod simple-prod-query-678d965c78-7q6sr.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:17 +0000 UTC Normal Pod simple-prod-query-678d965c78-7q6sr.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:18 +0000 UTC Warning Pod simple-prod-query-678d965c78-7q6sr.spec.containers{jaeger-query} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:19 +0000 UTC Warning Pod simple-prod-collector-75f468dd87-t9crf.spec.containers{jaeger-collector} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:25 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestlearningsawfishsimpleprod-1-97bf5p7n.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:29 +0000 UTC Warning Pod simple-prod-collector-75f468dd87-t9crf.spec.containers{jaeger-collector} BackOff Back-off restarting failed container jaeger-collector in pod simple-prod-collector-75f468dd87-t9crf_kuttl-test-learning-sawfish(68978b26-4a65-43aa-9e33-f926f36c4846) kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:29 +0000 UTC Warning Pod simple-prod-query-678d965c78-7q6sr.spec.containers{jaeger-query} BackOff Back-off restarting failed container jaeger-query in pod simple-prod-query-678d965c78-7q6sr_kuttl-test-learning-sawfish(b9297523-8a82-4c0e-a04f-a4f3ac2abb48) kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:30 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestlearningsawfishsimpleprod-1-97bf5p7n.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:30 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:30 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:30 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:50 +0000 UTC Normal Pod simple-prod-query-678d965c78-7q6sr.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:50 +0000 UTC Normal ReplicaSet.apps simple-prod-query-678d965c78 SuccessfulDelete Deleted pod: simple-prod-query-678d965c78-7q6sr replicaset-controller logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:50 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-678d965c78 to 0 from 1 deployment-controller logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:51 +0000 UTC Normal Pod simple-prod-query-b76766dc6-l7spm Binding Scheduled Successfully assigned kuttl-test-learning-sawfish/simple-prod-query-b76766dc6-l7spm to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:51 +0000 UTC Normal Pod simple-prod-query-b76766dc6-l7spm AddedInterface Add eth0 [10.131.0.73/23] from ovn-kubernetes multus logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:51 +0000 UTC Normal Pod simple-prod-query-b76766dc6-l7spm.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:51 +0000 UTC Normal ReplicaSet.apps simple-prod-query-b76766dc6 SuccessfulCreate Created pod: simple-prod-query-b76766dc6-l7spm replicaset-controller logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:51 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-b76766dc6 to 1 deployment-controller logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:52 +0000 UTC Normal Pod simple-prod-query-b76766dc6-l7spm.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:52 +0000 UTC Normal Pod simple-prod-query-b76766dc6-l7spm.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:52 +0000 UTC Normal Pod simple-prod-query-b76766dc6-l7spm.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:52 +0000 UTC Normal Pod simple-prod-query-b76766dc6-l7spm.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:52 +0000 UTC Normal Pod simple-prod-query-b76766dc6-l7spm.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:52 +0000 UTC Normal Pod simple-prod-query-b76766dc6-l7spm.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:52 +0000 UTC Normal Pod simple-prod-query-b76766dc6-l7spm.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:52 +0000 UTC Normal Pod simple-prod-query-b76766dc6-l7spm.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:55 +0000 UTC Normal Pod report-span-6jqkr Binding Scheduled Successfully assigned kuttl-test-learning-sawfish/report-span-6jqkr to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:55 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-6jqkr job-controller logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:56 +0000 UTC Normal Pod check-span-56lp9 Binding Scheduled Successfully assigned kuttl-test-learning-sawfish/check-span-56lp9 to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:56 +0000 UTC Normal Pod check-span-56lp9 AddedInterface Add eth0 [10.128.2.37/23] from ovn-kubernetes multus logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:56 +0000 UTC Normal Pod check-span-56lp9.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:56 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-56lp9 job-controller logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:56 +0000 UTC Normal Pod report-span-6jqkr AddedInterface Add eth0 [10.128.2.36/23] from ovn-kubernetes multus logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:56 +0000 UTC Normal Pod report-span-6jqkr.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:57 +0000 UTC Normal Pod check-span-56lp9.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 221ms (221ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:57 +0000 UTC Normal Pod check-span-56lp9.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:57 +0000 UTC Normal Pod check-span-56lp9.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:57 +0000 UTC Normal Pod report-span-6jqkr.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 378ms (378ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:57 +0000 UTC Normal Pod report-span-6jqkr.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:23:57 +0000 UTC Normal Pod report-span-6jqkr.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 18:24:08 | examples-simple-prod | 2024-09-30 18:24:08 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 18:24:08 | examples-simple-prod | Deleting namespace: kuttl-test-learning-sawfish === CONT kuttl/harness/examples-with-sampling logger.go:42: 18:24:20 | examples-with-sampling | Creating namespace: kuttl-test-sunny-bee logger.go:42: 18:24:20 | examples-with-sampling/0-install | starting test step 0-install logger.go:42: 18:24:20 | examples-with-sampling/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 18:24:20 | examples-with-sampling/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 18:24:20 | examples-with-sampling/0-install | >>>> Creating namespace kuttl-test-sunny-bee logger.go:42: 18:24:20 | examples-with-sampling/0-install | kubectl create namespace kuttl-test-sunny-bee 2>&1 | grep -v "already exists" || true logger.go:42: 18:24:21 | examples-with-sampling/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-sunny-bee 2>&1 | grep -v "already exists" || true logger.go:42: 18:24:21 | examples-with-sampling/0-install | service/cassandra created logger.go:42: 18:24:21 | examples-with-sampling/0-install | statefulset.apps/cassandra created logger.go:42: 18:24:21 | examples-with-sampling/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 18:24:31 | examples-with-sampling/0-install | test step completed 0-install logger.go:42: 18:24:31 | examples-with-sampling/1-install | starting test step 1-install logger.go:42: 18:24:31 | examples-with-sampling/1-install | Jaeger:kuttl-test-sunny-bee/with-sampling created logger.go:42: 18:24:38 | examples-with-sampling/1-install | test step completed 1-install logger.go:42: 18:24:38 | examples-with-sampling/2-smoke-test | starting test step 2-smoke-test logger.go:42: 18:24:38 | examples-with-sampling/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-sampling /dev/null] logger.go:42: 18:24:40 | examples-with-sampling/2-smoke-test | Warning: resource jaegers/with-sampling is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 18:24:46 | examples-with-sampling/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 18:24:47 | examples-with-sampling/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 18:24:47 | examples-with-sampling/2-smoke-test | job.batch/report-span created logger.go:42: 18:24:47 | examples-with-sampling/2-smoke-test | job.batch/check-span created logger.go:42: 18:24:59 | examples-with-sampling/2-smoke-test | test step completed 2-smoke-test logger.go:42: 18:24:59 | examples-with-sampling/3- | starting test step 3- logger.go:42: 18:24:59 | examples-with-sampling/3- | test step completed 3- logger.go:42: 18:24:59 | examples-with-sampling | examples-with-sampling events from ns kuttl-test-sunny-bee: logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:21 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-sunny-bee/cassandra-0 to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:21 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:22 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.128.2.38/23] from ovn-kubernetes multus logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:22 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:25 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 3.825s (3.825s including waiting). Image size: 305863309 bytes. kubelet logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:26 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:26 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:26 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-sunny-bee/cassandra-1 to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:26 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.131.0.74/23] from ovn-kubernetes multus logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:26 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:26 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:30 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 3.677s (3.677s including waiting). Image size: 305863309 bytes. kubelet logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:30 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:30 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:35 +0000 UTC Normal Pod with-sampling-9b6d7b9cc-hv55v Binding Scheduled Successfully assigned kuttl-test-sunny-bee/with-sampling-9b6d7b9cc-hv55v to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:35 +0000 UTC Normal ReplicaSet.apps with-sampling-9b6d7b9cc SuccessfulCreate Created pod: with-sampling-9b6d7b9cc-hv55v replicaset-controller logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:35 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-9b6d7b9cc to 1 deployment-controller logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:36 +0000 UTC Normal Pod with-sampling-9b6d7b9cc-hv55v AddedInterface Add eth0 [10.129.2.70/23] from ovn-kubernetes multus logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:36 +0000 UTC Normal Pod with-sampling-9b6d7b9cc-hv55v.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:36 +0000 UTC Normal Pod with-sampling-9b6d7b9cc-hv55v.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:36 +0000 UTC Normal Pod with-sampling-9b6d7b9cc-hv55v.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:36 +0000 UTC Normal Pod with-sampling-9b6d7b9cc-hv55v.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:36 +0000 UTC Normal Pod with-sampling-9b6d7b9cc-hv55v.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:36 +0000 UTC Normal Pod with-sampling-9b6d7b9cc-hv55v.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:41 +0000 UTC Normal Pod with-sampling-9b6d7b9cc-hv55v.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:41 +0000 UTC Normal Pod with-sampling-9b6d7b9cc-hv55v.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:41 +0000 UTC Normal ReplicaSet.apps with-sampling-9b6d7b9cc SuccessfulDelete Deleted pod: with-sampling-9b6d7b9cc-hv55v replicaset-controller logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:41 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled down replica set with-sampling-9b6d7b9cc to 0 from 1 deployment-controller logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:42 +0000 UTC Normal Pod with-sampling-674dbc747b-gxtcf Binding Scheduled Successfully assigned kuttl-test-sunny-bee/with-sampling-674dbc747b-gxtcf to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:42 +0000 UTC Normal ReplicaSet.apps with-sampling-674dbc747b SuccessfulCreate Created pod: with-sampling-674dbc747b-gxtcf replicaset-controller logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:42 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-674dbc747b to 1 deployment-controller logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:43 +0000 UTC Normal Pod with-sampling-674dbc747b-gxtcf AddedInterface Add eth0 [10.129.2.71/23] from ovn-kubernetes multus logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:43 +0000 UTC Normal Pod with-sampling-674dbc747b-gxtcf.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:43 +0000 UTC Normal Pod with-sampling-674dbc747b-gxtcf.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:43 +0000 UTC Normal Pod with-sampling-674dbc747b-gxtcf.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:43 +0000 UTC Normal Pod with-sampling-674dbc747b-gxtcf.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:43 +0000 UTC Normal Pod with-sampling-674dbc747b-gxtcf.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:43 +0000 UTC Normal Pod with-sampling-674dbc747b-gxtcf.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:44 +0000 UTC Warning Pod with-sampling-674dbc747b-gxtcf FailedMount MountVolume.SetUp failed for volume "with-sampling-service-ca" : configmap references non-existent config key: service-ca.crt kubelet logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:47 +0000 UTC Normal Pod check-span-zn2j6 Binding Scheduled Successfully assigned kuttl-test-sunny-bee/check-span-zn2j6 to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:47 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-zn2j6 job-controller logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:47 +0000 UTC Normal Pod report-span-r7x6b Binding Scheduled Successfully assigned kuttl-test-sunny-bee/report-span-r7x6b to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:47 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-r7x6b job-controller logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:48 +0000 UTC Normal Pod check-span-zn2j6 AddedInterface Add eth0 [10.128.2.39/23] from ovn-kubernetes multus logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:48 +0000 UTC Normal Pod check-span-zn2j6.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:48 +0000 UTC Normal Pod check-span-zn2j6.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 207ms (207ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:48 +0000 UTC Normal Pod check-span-zn2j6.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:48 +0000 UTC Normal Pod check-span-zn2j6.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:48 +0000 UTC Normal Pod report-span-r7x6b AddedInterface Add eth0 [10.131.0.75/23] from ovn-kubernetes multus logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:48 +0000 UTC Normal Pod report-span-r7x6b.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:48 +0000 UTC Normal Pod report-span-r7x6b.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 279ms (279ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:48 +0000 UTC Normal Pod report-span-r7x6b.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:48 +0000 UTC Normal Pod report-span-r7x6b.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 18:24:59 | examples-with-sampling | 2024-09-30 18:24:59 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 18:24:59 | examples-with-sampling | Deleting namespace: kuttl-test-sunny-bee === CONT kuttl/harness/examples-with-cassandra logger.go:42: 18:25:17 | examples-with-cassandra | Creating namespace: kuttl-test-handy-arachnid logger.go:42: 18:25:17 | examples-with-cassandra/0-install | starting test step 0-install logger.go:42: 18:25:17 | examples-with-cassandra/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 18:25:17 | examples-with-cassandra/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 18:25:17 | examples-with-cassandra/0-install | >>>> Creating namespace kuttl-test-handy-arachnid logger.go:42: 18:25:17 | examples-with-cassandra/0-install | kubectl create namespace kuttl-test-handy-arachnid 2>&1 | grep -v "already exists" || true logger.go:42: 18:25:17 | examples-with-cassandra/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-handy-arachnid 2>&1 | grep -v "already exists" || true logger.go:42: 18:25:18 | examples-with-cassandra/0-install | service/cassandra created logger.go:42: 18:25:18 | examples-with-cassandra/0-install | statefulset.apps/cassandra created logger.go:42: 18:25:18 | examples-with-cassandra/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 18:25:21 | examples-with-cassandra/0-install | test step completed 0-install logger.go:42: 18:25:21 | examples-with-cassandra/1-install | starting test step 1-install logger.go:42: 18:25:21 | examples-with-cassandra/1-install | Jaeger:kuttl-test-handy-arachnid/with-cassandra created logger.go:42: 18:25:38 | examples-with-cassandra/1-install | test step completed 1-install logger.go:42: 18:25:38 | examples-with-cassandra/2-smoke-test | starting test step 2-smoke-test logger.go:42: 18:25:38 | examples-with-cassandra/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-cassandra /dev/null] logger.go:42: 18:25:40 | examples-with-cassandra/2-smoke-test | Warning: resource jaegers/with-cassandra is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 18:25:47 | examples-with-cassandra/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 18:25:47 | examples-with-cassandra/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 18:25:47 | examples-with-cassandra/2-smoke-test | job.batch/report-span created logger.go:42: 18:25:47 | examples-with-cassandra/2-smoke-test | job.batch/check-span created logger.go:42: 18:26:00 | examples-with-cassandra/2-smoke-test | test step completed 2-smoke-test logger.go:42: 18:26:00 | examples-with-cassandra | examples-with-cassandra events from ns kuttl-test-handy-arachnid: logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:18 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-handy-arachnid/cassandra-0 to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:18 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.128.2.40/23] from ovn-kubernetes multus logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:18 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:19 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:19 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:19 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:19 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-handy-arachnid/cassandra-1 to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:19 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.131.0.76/23] from ovn-kubernetes multus logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:19 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:19 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:20 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:20 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:25 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-k28zk Binding Scheduled Successfully assigned kuttl-test-handy-arachnid/with-cassandra-cassandra-schema-job-k28zk to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:25 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job SuccessfulCreate Created pod: with-cassandra-cassandra-schema-job-k28zk job-controller logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:26 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-k28zk AddedInterface Add eth0 [10.129.2.72/23] from ovn-kubernetes multus logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:26 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-k28zk.spec.containers{with-cassandra-cassandra-schema-job} Pulling Pulling image "jaegertracing/jaeger-cassandra-schema:1.57.0" kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:30 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-k28zk.spec.containers{with-cassandra-cassandra-schema-job} Pulled Successfully pulled image "jaegertracing/jaeger-cassandra-schema:1.57.0" in 4.529s (4.529s including waiting). Image size: 381891613 bytes. kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:30 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-k28zk.spec.containers{with-cassandra-cassandra-schema-job} Created Created container with-cassandra-cassandra-schema-job kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:30 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-k28zk.spec.containers{with-cassandra-cassandra-schema-job} Started Started container with-cassandra-cassandra-schema-job kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:36 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job Completed Job completed job-controller logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:37 +0000 UTC Normal Pod with-cassandra-6d994fb7cd-x8lq7 Binding Scheduled Successfully assigned kuttl-test-handy-arachnid/with-cassandra-6d994fb7cd-x8lq7 to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:37 +0000 UTC Normal Pod with-cassandra-6d994fb7cd-x8lq7 AddedInterface Add eth0 [10.129.2.73/23] from ovn-kubernetes multus logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:37 +0000 UTC Normal Pod with-cassandra-6d994fb7cd-x8lq7.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:37 +0000 UTC Normal Pod with-cassandra-6d994fb7cd-x8lq7.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:37 +0000 UTC Normal ReplicaSet.apps with-cassandra-6d994fb7cd SuccessfulCreate Created pod: with-cassandra-6d994fb7cd-x8lq7 replicaset-controller logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:37 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-6d994fb7cd to 1 deployment-controller logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:38 +0000 UTC Normal Pod with-cassandra-6d994fb7cd-x8lq7.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:38 +0000 UTC Normal Pod with-cassandra-6d994fb7cd-x8lq7.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:38 +0000 UTC Normal Pod with-cassandra-6d994fb7cd-x8lq7.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:38 +0000 UTC Normal Pod with-cassandra-6d994fb7cd-x8lq7.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:42 +0000 UTC Normal Pod with-cassandra-6d994fb7cd-x8lq7.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:42 +0000 UTC Normal Pod with-cassandra-6d994fb7cd-x8lq7.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:42 +0000 UTC Normal ReplicaSet.apps with-cassandra-6d994fb7cd SuccessfulDelete Deleted pod: with-cassandra-6d994fb7cd-x8lq7 replicaset-controller logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:42 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled down replica set with-cassandra-6d994fb7cd to 0 from 1 deployment-controller logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:43 +0000 UTC Normal Pod with-cassandra-6d7d67958d-lq8rf Binding Scheduled Successfully assigned kuttl-test-handy-arachnid/with-cassandra-6d7d67958d-lq8rf to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:43 +0000 UTC Normal ReplicaSet.apps with-cassandra-6d7d67958d SuccessfulCreate Created pod: with-cassandra-6d7d67958d-lq8rf replicaset-controller logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:43 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-6d7d67958d to 1 deployment-controller logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:44 +0000 UTC Normal Pod with-cassandra-6d7d67958d-lq8rf AddedInterface Add eth0 [10.129.2.74/23] from ovn-kubernetes multus logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:44 +0000 UTC Normal Pod with-cassandra-6d7d67958d-lq8rf.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:44 +0000 UTC Normal Pod with-cassandra-6d7d67958d-lq8rf.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:44 +0000 UTC Normal Pod with-cassandra-6d7d67958d-lq8rf.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:44 +0000 UTC Normal Pod with-cassandra-6d7d67958d-lq8rf.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:44 +0000 UTC Normal Pod with-cassandra-6d7d67958d-lq8rf.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:44 +0000 UTC Normal Pod with-cassandra-6d7d67958d-lq8rf.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:47 +0000 UTC Normal Pod check-span-8pz2z Binding Scheduled Successfully assigned kuttl-test-handy-arachnid/check-span-8pz2z to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:47 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-8pz2z job-controller logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:47 +0000 UTC Normal Pod report-span-gpzpn Binding Scheduled Successfully assigned kuttl-test-handy-arachnid/report-span-gpzpn to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:47 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-gpzpn job-controller logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:48 +0000 UTC Normal Pod check-span-8pz2z AddedInterface Add eth0 [10.131.0.77/23] from ovn-kubernetes multus logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:48 +0000 UTC Normal Pod check-span-8pz2z.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:48 +0000 UTC Normal Pod check-span-8pz2z.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 211ms (211ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:48 +0000 UTC Normal Pod check-span-8pz2z.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:48 +0000 UTC Normal Pod check-span-8pz2z.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:48 +0000 UTC Normal Pod report-span-gpzpn AddedInterface Add eth0 [10.128.2.41/23] from ovn-kubernetes multus logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:48 +0000 UTC Normal Pod report-span-gpzpn.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:48 +0000 UTC Normal Pod report-span-gpzpn.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 263ms (263ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:48 +0000 UTC Normal Pod report-span-gpzpn.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:48 +0000 UTC Normal Pod report-span-gpzpn.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 18:26:00 | examples-with-cassandra | 2024-09-30 18:25:59 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 18:26:00 | examples-with-cassandra | Deleting namespace: kuttl-test-handy-arachnid === CONT kuttl/harness/examples-with-badger logger.go:42: 18:26:12 | examples-with-badger | Creating namespace: kuttl-test-accepted-bass logger.go:42: 18:26:12 | examples-with-badger/0-install | starting test step 0-install logger.go:42: 18:26:12 | examples-with-badger/0-install | Jaeger:kuttl-test-accepted-bass/with-badger created logger.go:42: 18:26:19 | examples-with-badger/0-install | test step completed 0-install logger.go:42: 18:26:19 | examples-with-badger/1-smoke-test | starting test step 1-smoke-test logger.go:42: 18:26:19 | examples-with-badger/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-badger /dev/null] logger.go:42: 18:26:21 | examples-with-badger/1-smoke-test | Warning: resource jaegers/with-badger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 18:26:27 | examples-with-badger/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 18:26:28 | examples-with-badger/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 18:26:28 | examples-with-badger/1-smoke-test | job.batch/report-span created logger.go:42: 18:26:28 | examples-with-badger/1-smoke-test | job.batch/check-span created logger.go:42: 18:26:40 | examples-with-badger/1-smoke-test | test step completed 1-smoke-test logger.go:42: 18:26:40 | examples-with-badger | examples-with-badger events from ns kuttl-test-accepted-bass: logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:16 +0000 UTC Normal Pod with-badger-7497fd4446-4799h Binding Scheduled Successfully assigned kuttl-test-accepted-bass/with-badger-7497fd4446-4799h to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:16 +0000 UTC Normal Pod with-badger-7497fd4446-4799h AddedInterface Add eth0 [10.131.0.78/23] from ovn-kubernetes multus logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:16 +0000 UTC Normal Pod with-badger-7497fd4446-4799h.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:16 +0000 UTC Normal ReplicaSet.apps with-badger-7497fd4446 SuccessfulCreate Created pod: with-badger-7497fd4446-4799h replicaset-controller logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:16 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-7497fd4446 to 1 deployment-controller logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:17 +0000 UTC Normal Pod with-badger-7497fd4446-4799h.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:17 +0000 UTC Normal Pod with-badger-7497fd4446-4799h.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:17 +0000 UTC Normal Pod with-badger-7497fd4446-4799h.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:17 +0000 UTC Normal Pod with-badger-7497fd4446-4799h.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:17 +0000 UTC Normal Pod with-badger-7497fd4446-4799h.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:22 +0000 UTC Normal Pod with-badger-7497fd4446-4799h.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:22 +0000 UTC Normal Pod with-badger-7497fd4446-4799h.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:22 +0000 UTC Normal ReplicaSet.apps with-badger-7497fd4446 SuccessfulDelete Deleted pod: with-badger-7497fd4446-4799h replicaset-controller logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:22 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled down replica set with-badger-7497fd4446 to 0 from 1 deployment-controller logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:23 +0000 UTC Normal Pod with-badger-5864b49bd6-gv2cq Binding Scheduled Successfully assigned kuttl-test-accepted-bass/with-badger-5864b49bd6-gv2cq to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:23 +0000 UTC Normal ReplicaSet.apps with-badger-5864b49bd6 SuccessfulCreate Created pod: with-badger-5864b49bd6-gv2cq replicaset-controller logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:23 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-5864b49bd6 to 1 deployment-controller logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:24 +0000 UTC Normal Pod with-badger-5864b49bd6-gv2cq AddedInterface Add eth0 [10.131.0.79/23] from ovn-kubernetes multus logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:24 +0000 UTC Normal Pod with-badger-5864b49bd6-gv2cq.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:24 +0000 UTC Normal Pod with-badger-5864b49bd6-gv2cq.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:24 +0000 UTC Normal Pod with-badger-5864b49bd6-gv2cq.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:24 +0000 UTC Normal Pod with-badger-5864b49bd6-gv2cq.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:24 +0000 UTC Normal Pod with-badger-5864b49bd6-gv2cq.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:24 +0000 UTC Normal Pod with-badger-5864b49bd6-gv2cq.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:28 +0000 UTC Normal Pod check-span-5vkg4 Binding Scheduled Successfully assigned kuttl-test-accepted-bass/check-span-5vkg4 to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:28 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-5vkg4 job-controller logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:28 +0000 UTC Normal Pod report-span-xmzgg Binding Scheduled Successfully assigned kuttl-test-accepted-bass/report-span-xmzgg to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:28 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-xmzgg job-controller logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:29 +0000 UTC Normal Pod check-span-5vkg4 AddedInterface Add eth0 [10.129.2.75/23] from ovn-kubernetes multus logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:29 +0000 UTC Normal Pod check-span-5vkg4.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:29 +0000 UTC Normal Pod check-span-5vkg4.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 246ms (246ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:29 +0000 UTC Normal Pod check-span-5vkg4.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:29 +0000 UTC Normal Pod check-span-5vkg4.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:29 +0000 UTC Normal Pod report-span-xmzgg AddedInterface Add eth0 [10.128.2.42/23] from ovn-kubernetes multus logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:29 +0000 UTC Normal Pod report-span-xmzgg.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:29 +0000 UTC Normal Pod report-span-xmzgg.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 257ms (257ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:29 +0000 UTC Normal Pod report-span-xmzgg.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:29 +0000 UTC Normal Pod report-span-xmzgg.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 18:26:40 | examples-with-badger | 2024-09-30 18:26:40 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 18:26:40 | examples-with-badger | Deleting namespace: kuttl-test-accepted-bass === CONT kuttl/harness/examples-simplest logger.go:42: 18:26:52 | examples-simplest | Creating namespace: kuttl-test-huge-martin logger.go:42: 18:26:52 | examples-simplest/0-install | starting test step 0-install logger.go:42: 18:26:53 | examples-simplest/0-install | Jaeger:kuttl-test-huge-martin/simplest created logger.go:42: 18:26:59 | examples-simplest/0-install | test step completed 0-install logger.go:42: 18:26:59 | examples-simplest/1-smoke-test | starting test step 1-smoke-test logger.go:42: 18:26:59 | examples-simplest/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 18:27:00 | examples-simplest/1-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 18:27:07 | examples-simplest/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 18:27:07 | examples-simplest/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 18:27:07 | examples-simplest/1-smoke-test | job.batch/report-span created logger.go:42: 18:27:07 | examples-simplest/1-smoke-test | job.batch/check-span created logger.go:42: 18:27:20 | examples-simplest/1-smoke-test | test step completed 1-smoke-test logger.go:42: 18:27:20 | examples-simplest | examples-simplest events from ns kuttl-test-huge-martin: logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:26:56 +0000 UTC Normal Pod simplest-844bf4948-4986n Binding Scheduled Successfully assigned kuttl-test-huge-martin/simplest-844bf4948-4986n to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:26:56 +0000 UTC Normal ReplicaSet.apps simplest-844bf4948 SuccessfulCreate Created pod: simplest-844bf4948-4986n replicaset-controller logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:26:56 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-844bf4948 to 1 deployment-controller logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:26:57 +0000 UTC Normal Pod simplest-844bf4948-4986n AddedInterface Add eth0 [10.131.0.80/23] from ovn-kubernetes multus logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:26:57 +0000 UTC Normal Pod simplest-844bf4948-4986n.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:26:57 +0000 UTC Normal Pod simplest-844bf4948-4986n.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:26:57 +0000 UTC Normal Pod simplest-844bf4948-4986n.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:26:57 +0000 UTC Normal Pod simplest-844bf4948-4986n.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:26:57 +0000 UTC Normal Pod simplest-844bf4948-4986n.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:26:57 +0000 UTC Normal Pod simplest-844bf4948-4986n.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:27:03 +0000 UTC Normal Pod simplest-844bf4948-4986n.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:27:03 +0000 UTC Normal Pod simplest-844bf4948-4986n.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:27:03 +0000 UTC Normal ReplicaSet.apps simplest-844bf4948 SuccessfulDelete Deleted pod: simplest-844bf4948-4986n replicaset-controller logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:27:03 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-844bf4948 to 0 from 1 deployment-controller logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:27:04 +0000 UTC Normal Pod simplest-95f998fb4-nrhb2 Binding Scheduled Successfully assigned kuttl-test-huge-martin/simplest-95f998fb4-nrhb2 to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:27:04 +0000 UTC Normal ReplicaSet.apps simplest-95f998fb4 SuccessfulCreate Created pod: simplest-95f998fb4-nrhb2 replicaset-controller logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:27:04 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-95f998fb4 to 1 deployment-controller logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:27:05 +0000 UTC Normal Pod simplest-95f998fb4-nrhb2 AddedInterface Add eth0 [10.131.0.81/23] from ovn-kubernetes multus logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:27:05 +0000 UTC Normal Pod simplest-95f998fb4-nrhb2.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:27:05 +0000 UTC Normal Pod simplest-95f998fb4-nrhb2.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:27:05 +0000 UTC Normal Pod simplest-95f998fb4-nrhb2.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:27:05 +0000 UTC Normal Pod simplest-95f998fb4-nrhb2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:27:05 +0000 UTC Normal Pod simplest-95f998fb4-nrhb2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:27:05 +0000 UTC Normal Pod simplest-95f998fb4-nrhb2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:27:07 +0000 UTC Normal Pod check-span-cdmhw Binding Scheduled Successfully assigned kuttl-test-huge-martin/check-span-cdmhw to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:27:07 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-cdmhw job-controller logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:27:07 +0000 UTC Normal Pod report-span-lv9ft Binding Scheduled Successfully assigned kuttl-test-huge-martin/report-span-lv9ft to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:27:07 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-lv9ft job-controller logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:27:08 +0000 UTC Normal Pod check-span-cdmhw AddedInterface Add eth0 [10.128.2.43/23] from ovn-kubernetes multus logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:27:08 +0000 UTC Normal Pod check-span-cdmhw.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:27:08 +0000 UTC Normal Pod check-span-cdmhw.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 235ms (235ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:27:08 +0000 UTC Normal Pod check-span-cdmhw.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:27:08 +0000 UTC Normal Pod check-span-cdmhw.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:27:08 +0000 UTC Normal Pod report-span-lv9ft AddedInterface Add eth0 [10.129.2.76/23] from ovn-kubernetes multus logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:27:08 +0000 UTC Normal Pod report-span-lv9ft.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:27:09 +0000 UTC Normal Pod report-span-lv9ft.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 198ms (198ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:27:09 +0000 UTC Normal Pod report-span-lv9ft.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:27:09 +0000 UTC Normal Pod report-span-lv9ft.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 18:27:20 | examples-simplest | 2024-09-30 18:27:19 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 18:27:20 | examples-simplest | Deleting namespace: kuttl-test-huge-martin === CONT kuttl/harness/examples-simple-prod-with-volumes logger.go:42: 18:27:32 | examples-simple-prod-with-volumes | Ignoring 03-check-volume.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 18:27:32 | examples-simple-prod-with-volumes | Creating namespace: kuttl-test-kind-falcon logger.go:42: 18:27:32 | examples-simple-prod-with-volumes/1-install | starting test step 1-install logger.go:42: 18:27:32 | examples-simple-prod-with-volumes/1-install | Jaeger:kuttl-test-kind-falcon/simple-prod created logger.go:42: 18:28:10 | examples-simple-prod-with-volumes/1-install | test step completed 1-install logger.go:42: 18:28:10 | examples-simple-prod-with-volumes/2-smoke-test | starting test step 2-smoke-test logger.go:42: 18:28:10 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 18:28:12 | examples-simple-prod-with-volumes/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 18:28:18 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 18:28:19 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 18:28:19 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/report-span created logger.go:42: 18:28:19 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/check-span created logger.go:42: 18:28:31 | examples-simple-prod-with-volumes/2-smoke-test | test step completed 2-smoke-test logger.go:42: 18:28:31 | examples-simple-prod-with-volumes/3-check-volume | starting test step 3-check-volume logger.go:42: 18:28:31 | examples-simple-prod-with-volumes/3-check-volume | running command: [sh -c kubectl exec $(kubectl get pods -n $NAMESPACE -l app=jaeger -l app.kubernetes.io/component=collector -o yaml | /tmp/jaeger-tests/bin/yq e '.items[0].metadata.name') -n $NAMESPACE -- ls /usr/share/elasticsearch/data] logger.go:42: 18:28:32 | examples-simple-prod-with-volumes/3-check-volume | test step completed 3-check-volume logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | examples-simple-prod-with-volumes events from ns kuttl-test-kind-falcon: logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:27:39 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestkindfalconsimpleprod-1-6bc75db7c4 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestkindfalconsimpleprod-1-6bc75db7rttvc replicaset-controller logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:27:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestkindfalconsimpleprod-1-6bc75db7rttvc Binding Scheduled Successfully assigned kuttl-test-kind-falcon/elasticsearch-cdm-kuttltestkindfalconsimpleprod-1-6bc75db7rttvc to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:27:39 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestkindfalconsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestkindfalconsimpleprod-1-6bc75db7c4 to 1 deployment-controller logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:27:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestkindfalconsimpleprod-1-6bc75db7rttvc AddedInterface Add eth0 [10.129.2.77/23] from ovn-kubernetes multus logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:27:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestkindfalconsimpleprod-1-6bc75db7rttvc.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:89629964e44058bafc90393a235b08c7c974f05513aecbfe7229134da732f2b5" already present on machine kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:27:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestkindfalconsimpleprod-1-6bc75db7rttvc.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:27:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestkindfalconsimpleprod-1-6bc75db7rttvc.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:27:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestkindfalconsimpleprod-1-6bc75db7rttvc.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:a8f53dec42a46c5bf8ac7f2888848c01e70f405d27211e2a269730c888929faf" already present on machine kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:27:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestkindfalconsimpleprod-1-6bc75db7rttvc.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:27:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestkindfalconsimpleprod-1-6bc75db7rttvc.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:27:50 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestkindfalconsimpleprod-1-6bc75db7rttvc.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:27:55 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestkindfalconsimpleprod-1-6bc75db7rttvc.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:07 +0000 UTC Normal Pod simple-prod-collector-577885f76c-fmxwk Binding Scheduled Successfully assigned kuttl-test-kind-falcon/simple-prod-collector-577885f76c-fmxwk to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:07 +0000 UTC Normal Pod simple-prod-collector-577885f76c-fmxwk AddedInterface Add eth0 [10.128.2.44/23] from ovn-kubernetes multus logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:07 +0000 UTC Normal Pod simple-prod-collector-577885f76c-fmxwk.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:07 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-577885f76c SuccessfulCreate Created pod: simple-prod-collector-577885f76c-fmxwk replicaset-controller logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:07 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-577885f76c to 1 deployment-controller logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:07 +0000 UTC Normal Pod simple-prod-query-55d864c4b-97h9t Binding Scheduled Successfully assigned kuttl-test-kind-falcon/simple-prod-query-55d864c4b-97h9t to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:07 +0000 UTC Normal ReplicaSet.apps simple-prod-query-55d864c4b SuccessfulCreate Created pod: simple-prod-query-55d864c4b-97h9t replicaset-controller logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:07 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-55d864c4b to 1 deployment-controller logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:08 +0000 UTC Normal Pod simple-prod-collector-577885f76c-fmxwk.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:08 +0000 UTC Normal Pod simple-prod-collector-577885f76c-fmxwk.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:08 +0000 UTC Normal Pod simple-prod-query-55d864c4b-97h9t AddedInterface Add eth0 [10.131.0.82/23] from ovn-kubernetes multus logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:08 +0000 UTC Normal Pod simple-prod-query-55d864c4b-97h9t.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:08 +0000 UTC Normal Pod simple-prod-query-55d864c4b-97h9t.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:08 +0000 UTC Normal Pod simple-prod-query-55d864c4b-97h9t.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:08 +0000 UTC Normal Pod simple-prod-query-55d864c4b-97h9t.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:08 +0000 UTC Normal Pod simple-prod-query-55d864c4b-97h9t.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:08 +0000 UTC Normal Pod simple-prod-query-55d864c4b-97h9t.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:08 +0000 UTC Normal Pod simple-prod-query-55d864c4b-97h9t.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:08 +0000 UTC Normal Pod simple-prod-query-55d864c4b-97h9t.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:08 +0000 UTC Normal Pod simple-prod-query-55d864c4b-97h9t.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:14 +0000 UTC Normal Pod simple-prod-query-55d864c4b-97h9t.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:14 +0000 UTC Normal Pod simple-prod-query-55d864c4b-97h9t.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:14 +0000 UTC Normal Pod simple-prod-query-55d864c4b-97h9t.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:14 +0000 UTC Normal ReplicaSet.apps simple-prod-query-55d864c4b SuccessfulDelete Deleted pod: simple-prod-query-55d864c4b-97h9t replicaset-controller logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:14 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-55d864c4b to 0 from 1 deployment-controller logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:15 +0000 UTC Normal Pod simple-prod-query-68c755f688-jqbx6 Binding Scheduled Successfully assigned kuttl-test-kind-falcon/simple-prod-query-68c755f688-jqbx6 to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:15 +0000 UTC Normal ReplicaSet.apps simple-prod-query-68c755f688 SuccessfulCreate Created pod: simple-prod-query-68c755f688-jqbx6 replicaset-controller logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:15 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-68c755f688 to 1 deployment-controller logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:16 +0000 UTC Normal Pod simple-prod-query-68c755f688-jqbx6 AddedInterface Add eth0 [10.131.0.83/23] from ovn-kubernetes multus logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:16 +0000 UTC Normal Pod simple-prod-query-68c755f688-jqbx6.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:16 +0000 UTC Normal Pod simple-prod-query-68c755f688-jqbx6.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:16 +0000 UTC Normal Pod simple-prod-query-68c755f688-jqbx6.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:16 +0000 UTC Normal Pod simple-prod-query-68c755f688-jqbx6.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:16 +0000 UTC Normal Pod simple-prod-query-68c755f688-jqbx6.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:16 +0000 UTC Normal Pod simple-prod-query-68c755f688-jqbx6.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:16 +0000 UTC Normal Pod simple-prod-query-68c755f688-jqbx6.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:16 +0000 UTC Normal Pod simple-prod-query-68c755f688-jqbx6.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:16 +0000 UTC Normal Pod simple-prod-query-68c755f688-jqbx6.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:19 +0000 UTC Normal Pod check-span-5fr79 Binding Scheduled Successfully assigned kuttl-test-kind-falcon/check-span-5fr79 to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:19 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-5fr79 job-controller logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:19 +0000 UTC Normal Pod report-span-f8k7q Binding Scheduled Successfully assigned kuttl-test-kind-falcon/report-span-f8k7q to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:19 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-f8k7q job-controller logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:20 +0000 UTC Normal Pod check-span-5fr79 AddedInterface Add eth0 [10.128.2.46/23] from ovn-kubernetes multus logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:20 +0000 UTC Normal Pod check-span-5fr79.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:20 +0000 UTC Normal Pod check-span-5fr79.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 248ms (248ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:20 +0000 UTC Normal Pod check-span-5fr79.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:20 +0000 UTC Normal Pod check-span-5fr79.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:20 +0000 UTC Normal Pod report-span-f8k7q AddedInterface Add eth0 [10.128.2.45/23] from ovn-kubernetes multus logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:20 +0000 UTC Normal Pod report-span-f8k7q.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:20 +0000 UTC Normal Pod report-span-f8k7q.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 215ms (215ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:20 +0000 UTC Normal Pod report-span-f8k7q.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:20 +0000 UTC Normal Pod report-span-f8k7q.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:25 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:25 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:25 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | 2024-09-30 18:28:31 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 18:28:32 | examples-simple-prod-with-volumes | Deleting namespace: kuttl-test-kind-falcon === CONT kuttl/harness/examples-business-application-injected-sidecar logger.go:42: 18:28:45 | examples-business-application-injected-sidecar | Creating namespace: kuttl-test-cuddly-seasnail logger.go:42: 18:28:45 | examples-business-application-injected-sidecar/0-install | starting test step 0-install logger.go:42: 18:28:45 | examples-business-application-injected-sidecar/0-install | Deployment:kuttl-test-cuddly-seasnail/myapp created logger.go:42: 18:28:45 | examples-business-application-injected-sidecar/0-install | test step completed 0-install logger.go:42: 18:28:45 | examples-business-application-injected-sidecar/1-install | starting test step 1-install logger.go:42: 18:28:45 | examples-business-application-injected-sidecar/1-install | Jaeger:kuttl-test-cuddly-seasnail/simplest created logger.go:42: 18:28:57 | examples-business-application-injected-sidecar/1-install | test step completed 1-install logger.go:42: 18:28:57 | examples-business-application-injected-sidecar/2-smoke-test | starting test step 2-smoke-test logger.go:42: 18:28:57 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 18:28:58 | examples-business-application-injected-sidecar/2-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 18:29:05 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 18:29:06 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 18:29:06 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/report-span created logger.go:42: 18:29:06 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/check-span created logger.go:42: 18:29:18 | examples-business-application-injected-sidecar/2-smoke-test | test step completed 2-smoke-test logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | examples-business-application-injected-sidecar events from ns kuttl-test-cuddly-seasnail: logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:45 +0000 UTC Normal Pod myapp-75f49f467-7ppn8 Binding Scheduled Successfully assigned kuttl-test-cuddly-seasnail/myapp-75f49f467-7ppn8 to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:45 +0000 UTC Normal ReplicaSet.apps myapp-75f49f467 SuccessfulCreate Created pod: myapp-75f49f467-7ppn8 replicaset-controller logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:45 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-75f49f467 to 1 deployment-controller logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:46 +0000 UTC Normal Pod myapp-75f49f467-7ppn8 AddedInterface Add eth0 [10.129.2.78/23] from ovn-kubernetes multus logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:46 +0000 UTC Normal Pod myapp-75f49f467-7ppn8.spec.containers{myapp} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:46 +0000 UTC Normal Pod myapp-75f49f467-7ppn8.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:46 +0000 UTC Normal Pod myapp-75f49f467-7ppn8.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:49 +0000 UTC Normal Pod myapp-5b9b9674fc-brbbr Binding Scheduled Successfully assigned kuttl-test-cuddly-seasnail/myapp-5b9b9674fc-brbbr to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:49 +0000 UTC Warning Pod myapp-5b9b9674fc-brbbr FailedMount MountVolume.SetUp failed for volume "simplest-trusted-ca" : configmap "simplest-trusted-ca" not found kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:49 +0000 UTC Warning Pod myapp-5b9b9674fc-brbbr FailedMount MountVolume.SetUp failed for volume "simplest-service-ca" : configmap "simplest-service-ca" not found kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:49 +0000 UTC Normal ReplicaSet.apps myapp-5b9b9674fc SuccessfulCreate Created pod: myapp-5b9b9674fc-brbbr replicaset-controller logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:49 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-5b9b9674fc to 1 deployment-controller logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:54 +0000 UTC Normal Pod simplest-644bdc9676-vvgjn Binding Scheduled Successfully assigned kuttl-test-cuddly-seasnail/simplest-644bdc9676-vvgjn to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:54 +0000 UTC Normal ReplicaSet.apps simplest-644bdc9676 SuccessfulCreate Created pod: simplest-644bdc9676-vvgjn replicaset-controller logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:54 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-644bdc9676 to 1 deployment-controller logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:55 +0000 UTC Normal Pod simplest-644bdc9676-vvgjn AddedInterface Add eth0 [10.131.0.84/23] from ovn-kubernetes multus logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:55 +0000 UTC Normal Pod simplest-644bdc9676-vvgjn.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:55 +0000 UTC Normal Pod simplest-644bdc9676-vvgjn.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:55 +0000 UTC Normal Pod simplest-644bdc9676-vvgjn.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:55 +0000 UTC Normal Pod simplest-644bdc9676-vvgjn.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:55 +0000 UTC Normal Pod simplest-644bdc9676-vvgjn.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:55 +0000 UTC Normal Pod simplest-644bdc9676-vvgjn.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:57 +0000 UTC Normal Pod myapp-5b9b9674fc-brbbr AddedInterface Add eth0 [10.128.2.47/23] from ovn-kubernetes multus logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:57 +0000 UTC Normal Pod myapp-5b9b9674fc-brbbr.spec.containers{myapp} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:57 +0000 UTC Normal Pod myapp-5b9b9674fc-brbbr.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:57 +0000 UTC Normal Pod myapp-5b9b9674fc-brbbr.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:57 +0000 UTC Normal Pod myapp-5b9b9674fc-brbbr.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:57 +0000 UTC Normal Pod myapp-5b9b9674fc-brbbr.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:57 +0000 UTC Normal Pod myapp-5b9b9674fc-brbbr.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:57 +0000 UTC Warning Pod myapp-75f49f467-7ppn8.spec.containers{myapp} Unhealthy Liveness probe failed: Get "http://10.129.2.78:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:58 +0000 UTC Normal Pod myapp-75f49f467-7ppn8.spec.containers{myapp} Killing Stopping container myapp kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:58 +0000 UTC Normal ReplicaSet.apps myapp-75f49f467 SuccessfulDelete Deleted pod: myapp-75f49f467-7ppn8 replicaset-controller logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:28:58 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled down replica set myapp-75f49f467 to 0 from 1 deployment-controller logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:29:00 +0000 UTC Normal Pod simplest-644bdc9676-vvgjn.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:29:00 +0000 UTC Normal Pod simplest-644bdc9676-vvgjn.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:29:00 +0000 UTC Normal ReplicaSet.apps simplest-644bdc9676 SuccessfulDelete Deleted pod: simplest-644bdc9676-vvgjn replicaset-controller logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:29:00 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-644bdc9676 to 0 from 1 deployment-controller logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:29:01 +0000 UTC Normal Pod simplest-75bf5d5fc-pz5j6 Binding Scheduled Successfully assigned kuttl-test-cuddly-seasnail/simplest-75bf5d5fc-pz5j6 to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:29:01 +0000 UTC Normal Pod simplest-75bf5d5fc-pz5j6 AddedInterface Add eth0 [10.131.0.85/23] from ovn-kubernetes multus logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:29:01 +0000 UTC Normal Pod simplest-75bf5d5fc-pz5j6.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:29:01 +0000 UTC Normal ReplicaSet.apps simplest-75bf5d5fc SuccessfulCreate Created pod: simplest-75bf5d5fc-pz5j6 replicaset-controller logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:29:01 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-75bf5d5fc to 1 deployment-controller logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:29:02 +0000 UTC Normal Pod simplest-75bf5d5fc-pz5j6.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:29:02 +0000 UTC Normal Pod simplest-75bf5d5fc-pz5j6.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:29:02 +0000 UTC Normal Pod simplest-75bf5d5fc-pz5j6.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:29:02 +0000 UTC Normal Pod simplest-75bf5d5fc-pz5j6.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:29:02 +0000 UTC Normal Pod simplest-75bf5d5fc-pz5j6.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:29:06 +0000 UTC Normal Pod check-span-t5qhb Binding Scheduled Successfully assigned kuttl-test-cuddly-seasnail/check-span-t5qhb to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:29:06 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-t5qhb job-controller logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:29:06 +0000 UTC Normal Pod report-span-8rdpn Binding Scheduled Successfully assigned kuttl-test-cuddly-seasnail/report-span-8rdpn to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:29:06 +0000 UTC Normal Pod report-span-8rdpn AddedInterface Add eth0 [10.129.2.79/23] from ovn-kubernetes multus logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:29:06 +0000 UTC Normal Pod report-span-8rdpn.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:29:06 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-8rdpn job-controller logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:29:07 +0000 UTC Normal Pod check-span-t5qhb AddedInterface Add eth0 [10.129.2.80/23] from ovn-kubernetes multus logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:29:07 +0000 UTC Normal Pod check-span-t5qhb.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:29:07 +0000 UTC Normal Pod check-span-t5qhb.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 204ms (204ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:29:07 +0000 UTC Normal Pod check-span-t5qhb.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:29:07 +0000 UTC Normal Pod check-span-t5qhb.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:29:07 +0000 UTC Normal Pod report-span-8rdpn.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 251ms (251ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:29:07 +0000 UTC Normal Pod report-span-8rdpn.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:29:07 +0000 UTC Normal Pod report-span-8rdpn.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:29:08 +0000 UTC Warning Pod myapp-5b9b9674fc-brbbr.spec.containers{myapp} Unhealthy Liveness probe failed: Get "http://10.128.2.47:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | 2024-09-30 18:29:17 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 18:29:18 | examples-business-application-injected-sidecar | Deleting namespace: kuttl-test-cuddly-seasnail === CONT kuttl/harness/examples-service-types logger.go:42: 18:29:30 | examples-service-types | Creating namespace: kuttl-test-unified-husky logger.go:42: 18:29:30 | examples-service-types/0-install | starting test step 0-install logger.go:42: 18:29:30 | examples-service-types/0-install | Jaeger:kuttl-test-unified-husky/service-types created logger.go:42: 18:29:38 | examples-service-types/0-install | test step completed 0-install logger.go:42: 18:29:38 | examples-service-types/1-smoke-test | starting test step 1-smoke-test logger.go:42: 18:29:38 | examples-service-types/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE service-types /dev/null] logger.go:42: 18:29:39 | examples-service-types/1-smoke-test | Warning: resource jaegers/service-types is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 18:29:46 | examples-service-types/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://service-types-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 18:29:46 | examples-service-types/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 18:29:46 | examples-service-types/1-smoke-test | job.batch/report-span created logger.go:42: 18:29:46 | examples-service-types/1-smoke-test | job.batch/check-span created logger.go:42: 18:29:59 | examples-service-types/1-smoke-test | test step completed 1-smoke-test logger.go:42: 18:29:59 | examples-service-types/2- | starting test step 2- logger.go:42: 18:29:59 | examples-service-types/2- | test step completed 2- logger.go:42: 18:29:59 | examples-service-types | examples-service-types events from ns kuttl-test-unified-husky: logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:34 +0000 UTC Normal Pod service-types-648c7bdbc4-snfcf Binding Scheduled Successfully assigned kuttl-test-unified-husky/service-types-648c7bdbc4-snfcf to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:34 +0000 UTC Warning Pod service-types-648c7bdbc4-snfcf FailedMount MountVolume.SetUp failed for volume "service-types-collector-tls-config-volume" : secret "service-types-collector-headless-tls" not found kubelet logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:34 +0000 UTC Normal ReplicaSet.apps service-types-648c7bdbc4 SuccessfulCreate Created pod: service-types-648c7bdbc4-snfcf replicaset-controller logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:34 +0000 UTC Normal Service service-types-collector EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:34 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-648c7bdbc4 to 1 deployment-controller logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:35 +0000 UTC Normal Pod service-types-648c7bdbc4-snfcf AddedInterface Add eth0 [10.131.0.86/23] from ovn-kubernetes multus logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:35 +0000 UTC Normal Pod service-types-648c7bdbc4-snfcf.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:36 +0000 UTC Normal Pod service-types-648c7bdbc4-snfcf.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:36 +0000 UTC Normal Pod service-types-648c7bdbc4-snfcf.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:36 +0000 UTC Normal Pod service-types-648c7bdbc4-snfcf.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:36 +0000 UTC Normal Pod service-types-648c7bdbc4-snfcf.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:36 +0000 UTC Normal Pod service-types-648c7bdbc4-snfcf.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:37 +0000 UTC Normal Service service-types-collector EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:37 +0000 UTC Normal Service service-types-query EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:40 +0000 UTC Normal Service service-types-query EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:41 +0000 UTC Normal Pod service-types-648c7bdbc4-snfcf.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:41 +0000 UTC Normal Pod service-types-648c7bdbc4-snfcf.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:41 +0000 UTC Normal ReplicaSet.apps service-types-648c7bdbc4 SuccessfulDelete Deleted pod: service-types-648c7bdbc4-snfcf replicaset-controller logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:41 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled down replica set service-types-648c7bdbc4 to 0 from 1 deployment-controller logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:42 +0000 UTC Normal Pod service-types-59468f4864-dl6b6 Binding Scheduled Successfully assigned kuttl-test-unified-husky/service-types-59468f4864-dl6b6 to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:42 +0000 UTC Normal Pod service-types-59468f4864-dl6b6 AddedInterface Add eth0 [10.131.0.87/23] from ovn-kubernetes multus logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:42 +0000 UTC Normal ReplicaSet.apps service-types-59468f4864 SuccessfulCreate Created pod: service-types-59468f4864-dl6b6 replicaset-controller logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:42 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-59468f4864 to 1 deployment-controller logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:43 +0000 UTC Normal Pod service-types-59468f4864-dl6b6.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:43 +0000 UTC Normal Pod service-types-59468f4864-dl6b6.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:43 +0000 UTC Normal Pod service-types-59468f4864-dl6b6.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:43 +0000 UTC Normal Pod service-types-59468f4864-dl6b6.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:43 +0000 UTC Normal Pod service-types-59468f4864-dl6b6.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:43 +0000 UTC Normal Pod service-types-59468f4864-dl6b6.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:46 +0000 UTC Normal Pod check-span-hjj6g Binding Scheduled Successfully assigned kuttl-test-unified-husky/check-span-hjj6g to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:46 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-hjj6g job-controller logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:46 +0000 UTC Normal Pod report-span-xmlkk Binding Scheduled Successfully assigned kuttl-test-unified-husky/report-span-xmlkk to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:46 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-xmlkk job-controller logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:47 +0000 UTC Normal Pod check-span-hjj6g AddedInterface Add eth0 [10.128.2.48/23] from ovn-kubernetes multus logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:47 +0000 UTC Normal Pod check-span-hjj6g.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:47 +0000 UTC Normal Pod check-span-hjj6g.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 238ms (238ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:47 +0000 UTC Normal Pod check-span-hjj6g.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:47 +0000 UTC Normal Pod check-span-hjj6g.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:47 +0000 UTC Normal Pod report-span-xmlkk AddedInterface Add eth0 [10.129.2.81/23] from ovn-kubernetes multus logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:47 +0000 UTC Normal Pod report-span-xmlkk.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:47 +0000 UTC Normal Pod report-span-xmlkk.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 240ms (240ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:47 +0000 UTC Normal Pod report-span-xmlkk.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:47 +0000 UTC Normal Pod report-span-xmlkk.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 18:29:59 | examples-service-types | 2024-09-30 18:29:58 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 18:29:59 | examples-service-types | Deleting namespace: kuttl-test-unified-husky === CONT kuttl/harness/examples-openshift-with-htpasswd logger.go:42: 18:30:11 | examples-openshift-with-htpasswd | Ignoring 00-install.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 18:30:11 | examples-openshift-with-htpasswd | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 18:30:11 | examples-openshift-with-htpasswd | Creating namespace: kuttl-test-relative-raptor logger.go:42: 18:30:11 | examples-openshift-with-htpasswd/0-install | starting test step 0-install logger.go:42: 18:30:11 | examples-openshift-with-htpasswd/0-install | Secret:kuttl-test-relative-raptor/htpasswd created logger.go:42: 18:30:11 | examples-openshift-with-htpasswd/0-install | test step completed 0-install logger.go:42: 18:30:11 | examples-openshift-with-htpasswd/1-install | starting test step 1-install logger.go:42: 18:30:11 | examples-openshift-with-htpasswd/1-install | Jaeger:kuttl-test-relative-raptor/with-htpasswd created logger.go:42: 18:30:19 | examples-openshift-with-htpasswd/1-install | test step completed 1-install logger.go:42: 18:30:19 | examples-openshift-with-htpasswd/2-check-unsecured | starting test step 2-check-unsecured logger.go:42: 18:30:19 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [./ensure-ingress-host.sh] logger.go:42: 18:30:19 | examples-openshift-with-htpasswd/2-check-unsecured | Checking the Ingress host value was populated logger.go:42: 18:30:19 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 0 logger.go:42: 18:30:19 | examples-openshift-with-htpasswd/2-check-unsecured | Hostname is with-htpasswd-kuttl-test-relative-raptor.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com logger.go:42: 18:30:19 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 18:30:19 | examples-openshift-with-htpasswd/2-check-unsecured | Checking an expected HTTP response logger.go:42: 18:30:19 | examples-openshift-with-htpasswd/2-check-unsecured | Running in OpenShift logger.go:42: 18:30:19 | examples-openshift-with-htpasswd/2-check-unsecured | Not using any secret logger.go:42: 18:30:19 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 1/30 the https://with-htpasswd-kuttl-test-relative-raptor.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com/search logger.go:42: 18:30:19 | examples-openshift-with-htpasswd/2-check-unsecured | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 18:30:19 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 2/30 the https://with-htpasswd-kuttl-test-relative-raptor.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com/search logger.go:42: 18:30:19 | examples-openshift-with-htpasswd/2-check-unsecured | HTTP response is 503. 403 expected. Waiting 10 s logger.go:42: 18:30:29 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 3/30 the https://with-htpasswd-kuttl-test-relative-raptor.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com/search logger.go:42: 18:30:29 | examples-openshift-with-htpasswd/2-check-unsecured | curl response asserted properly logger.go:42: 18:30:29 | examples-openshift-with-htpasswd/2-check-unsecured | test step completed 2-check-unsecured logger.go:42: 18:30:29 | examples-openshift-with-htpasswd/3-check-unauthorized | starting test step 3-check-unauthorized logger.go:42: 18:30:29 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [./ensure-ingress-host.sh] logger.go:42: 18:30:29 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking the Ingress host value was populated logger.go:42: 18:30:29 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 0 logger.go:42: 18:30:29 | examples-openshift-with-htpasswd/3-check-unauthorized | Hostname is with-htpasswd-kuttl-test-relative-raptor.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com logger.go:42: 18:30:29 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [sh -c JAEGER_USERNAME=wronguser JAEGER_PASSWORD=wrongpassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 18:30:29 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking an expected HTTP response logger.go:42: 18:30:29 | examples-openshift-with-htpasswd/3-check-unauthorized | Running in OpenShift logger.go:42: 18:30:29 | examples-openshift-with-htpasswd/3-check-unauthorized | Using Jaeger basic authentication logger.go:42: 18:30:29 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 1/30 the https://with-htpasswd-kuttl-test-relative-raptor.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com/search logger.go:42: 18:30:30 | examples-openshift-with-htpasswd/3-check-unauthorized | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 18:30:30 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 2/30 the https://with-htpasswd-kuttl-test-relative-raptor.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com/search logger.go:42: 18:30:30 | examples-openshift-with-htpasswd/3-check-unauthorized | curl response asserted properly logger.go:42: 18:30:30 | examples-openshift-with-htpasswd/3-check-unauthorized | test step completed 3-check-unauthorized logger.go:42: 18:30:30 | examples-openshift-with-htpasswd/4-check-authorized | starting test step 4-check-authorized logger.go:42: 18:30:30 | examples-openshift-with-htpasswd/4-check-authorized | running command: [./ensure-ingress-host.sh] logger.go:42: 18:30:30 | examples-openshift-with-htpasswd/4-check-authorized | Checking the Ingress host value was populated logger.go:42: 18:30:30 | examples-openshift-with-htpasswd/4-check-authorized | Try number 0 logger.go:42: 18:30:30 | examples-openshift-with-htpasswd/4-check-authorized | Hostname is with-htpasswd-kuttl-test-relative-raptor.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com logger.go:42: 18:30:30 | examples-openshift-with-htpasswd/4-check-authorized | running command: [sh -c JAEGER_USERNAME=awesomeuser JAEGER_PASSWORD=awesomepassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE with-htpasswd] logger.go:42: 18:30:30 | examples-openshift-with-htpasswd/4-check-authorized | Checking an expected HTTP response logger.go:42: 18:30:30 | examples-openshift-with-htpasswd/4-check-authorized | Running in OpenShift logger.go:42: 18:30:30 | examples-openshift-with-htpasswd/4-check-authorized | Using Jaeger basic authentication logger.go:42: 18:30:30 | examples-openshift-with-htpasswd/4-check-authorized | Try number 1/30 the https://with-htpasswd-kuttl-test-relative-raptor.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com/search logger.go:42: 18:30:30 | examples-openshift-with-htpasswd/4-check-authorized | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 18:30:30 | examples-openshift-with-htpasswd/4-check-authorized | Try number 2/30 the https://with-htpasswd-kuttl-test-relative-raptor.apps.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com/search logger.go:42: 18:30:30 | examples-openshift-with-htpasswd/4-check-authorized | curl response asserted properly logger.go:42: 18:30:30 | examples-openshift-with-htpasswd/4-check-authorized | test step completed 4-check-authorized logger.go:42: 18:30:30 | examples-openshift-with-htpasswd | examples-openshift-with-htpasswd events from ns kuttl-test-relative-raptor: logger.go:42: 18:30:30 | examples-openshift-with-htpasswd | 2024-09-30 18:30:15 +0000 UTC Normal Pod with-htpasswd-947967cbb-55srd Binding Scheduled Successfully assigned kuttl-test-relative-raptor/with-htpasswd-947967cbb-55srd to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:30:30 | examples-openshift-with-htpasswd | 2024-09-30 18:30:15 +0000 UTC Normal ReplicaSet.apps with-htpasswd-947967cbb SuccessfulCreate Created pod: with-htpasswd-947967cbb-55srd replicaset-controller logger.go:42: 18:30:30 | examples-openshift-with-htpasswd | 2024-09-30 18:30:15 +0000 UTC Normal Deployment.apps with-htpasswd ScalingReplicaSet Scaled up replica set with-htpasswd-947967cbb to 1 deployment-controller logger.go:42: 18:30:30 | examples-openshift-with-htpasswd | 2024-09-30 18:30:16 +0000 UTC Normal Pod with-htpasswd-947967cbb-55srd AddedInterface Add eth0 [10.131.0.89/23] from ovn-kubernetes multus logger.go:42: 18:30:30 | examples-openshift-with-htpasswd | 2024-09-30 18:30:16 +0000 UTC Normal Pod with-htpasswd-947967cbb-55srd.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 18:30:30 | examples-openshift-with-htpasswd | 2024-09-30 18:30:16 +0000 UTC Normal Pod with-htpasswd-947967cbb-55srd.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 18:30:30 | examples-openshift-with-htpasswd | 2024-09-30 18:30:16 +0000 UTC Normal Pod with-htpasswd-947967cbb-55srd.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 18:30:30 | examples-openshift-with-htpasswd | 2024-09-30 18:30:16 +0000 UTC Normal Pod with-htpasswd-947967cbb-55srd.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:30:30 | examples-openshift-with-htpasswd | 2024-09-30 18:30:16 +0000 UTC Normal Pod with-htpasswd-947967cbb-55srd.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:30:30 | examples-openshift-with-htpasswd | 2024-09-30 18:30:16 +0000 UTC Normal Pod with-htpasswd-947967cbb-55srd.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:30:30 | examples-openshift-with-htpasswd | Deleting namespace: kuttl-test-relative-raptor === CONT kuttl/harness/examples-all-in-one-with-options logger.go:42: 18:30:36 | examples-all-in-one-with-options | Creating namespace: kuttl-test-apt-mallard logger.go:42: 18:30:36 | examples-all-in-one-with-options/0-install | starting test step 0-install logger.go:42: 18:30:36 | examples-all-in-one-with-options/0-install | Jaeger:kuttl-test-apt-mallard/my-jaeger created logger.go:42: 18:30:42 | examples-all-in-one-with-options/0-install | test step completed 0-install logger.go:42: 18:30:42 | examples-all-in-one-with-options/1-smoke-test | starting test step 1-smoke-test logger.go:42: 18:30:42 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 18:30:44 | examples-all-in-one-with-options/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 18:30:50 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443/jaeger MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 18:30:51 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 18:30:51 | examples-all-in-one-with-options/1-smoke-test | job.batch/report-span created logger.go:42: 18:30:51 | examples-all-in-one-with-options/1-smoke-test | job.batch/check-span created logger.go:42: 18:31:04 | examples-all-in-one-with-options/1-smoke-test | test step completed 1-smoke-test logger.go:42: 18:31:04 | examples-all-in-one-with-options | examples-all-in-one-with-options events from ns kuttl-test-apt-mallard: logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:40 +0000 UTC Normal Pod my-jaeger-7b79f6d7c9-qbmm2 Binding Scheduled Successfully assigned kuttl-test-apt-mallard/my-jaeger-7b79f6d7c9-qbmm2 to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:40 +0000 UTC Normal ReplicaSet.apps my-jaeger-7b79f6d7c9 SuccessfulCreate Created pod: my-jaeger-7b79f6d7c9-qbmm2 replicaset-controller logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:40 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-7b79f6d7c9 to 1 deployment-controller logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:41 +0000 UTC Normal Pod my-jaeger-7b79f6d7c9-qbmm2 AddedInterface Add eth0 [10.131.0.90/23] from ovn-kubernetes multus logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:41 +0000 UTC Normal Pod my-jaeger-7b79f6d7c9-qbmm2.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:41 +0000 UTC Normal Pod my-jaeger-7b79f6d7c9-qbmm2.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:41 +0000 UTC Normal Pod my-jaeger-7b79f6d7c9-qbmm2.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:41 +0000 UTC Normal Pod my-jaeger-7b79f6d7c9-qbmm2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:41 +0000 UTC Normal Pod my-jaeger-7b79f6d7c9-qbmm2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:41 +0000 UTC Normal Pod my-jaeger-7b79f6d7c9-qbmm2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:47 +0000 UTC Normal Pod my-jaeger-7b79f6d7c9-qbmm2.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:47 +0000 UTC Normal Pod my-jaeger-7b79f6d7c9-qbmm2.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:47 +0000 UTC Normal ReplicaSet.apps my-jaeger-7b79f6d7c9 SuccessfulDelete Deleted pod: my-jaeger-7b79f6d7c9-qbmm2 replicaset-controller logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:47 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-7b79f6d7c9 to 0 from 1 deployment-controller logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:48 +0000 UTC Normal Pod my-jaeger-7c97d75-52bhr Binding Scheduled Successfully assigned kuttl-test-apt-mallard/my-jaeger-7c97d75-52bhr to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:48 +0000 UTC Normal ReplicaSet.apps my-jaeger-7c97d75 SuccessfulCreate Created pod: my-jaeger-7c97d75-52bhr replicaset-controller logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:48 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-7c97d75 to 1 deployment-controller logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:49 +0000 UTC Normal Pod my-jaeger-7c97d75-52bhr AddedInterface Add eth0 [10.131.0.91/23] from ovn-kubernetes multus logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:49 +0000 UTC Normal Pod my-jaeger-7c97d75-52bhr.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:49 +0000 UTC Normal Pod my-jaeger-7c97d75-52bhr.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:49 +0000 UTC Normal Pod my-jaeger-7c97d75-52bhr.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:49 +0000 UTC Normal Pod my-jaeger-7c97d75-52bhr.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:49 +0000 UTC Normal Pod my-jaeger-7c97d75-52bhr.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:49 +0000 UTC Normal Pod my-jaeger-7c97d75-52bhr.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:51 +0000 UTC Normal Pod check-span-7vtkt Binding Scheduled Successfully assigned kuttl-test-apt-mallard/check-span-7vtkt to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:51 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-7vtkt job-controller logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:51 +0000 UTC Normal Pod report-span-rbmzv Binding Scheduled Successfully assigned kuttl-test-apt-mallard/report-span-rbmzv to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:51 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-rbmzv job-controller logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:52 +0000 UTC Normal Pod check-span-7vtkt AddedInterface Add eth0 [10.128.2.49/23] from ovn-kubernetes multus logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:52 +0000 UTC Normal Pod check-span-7vtkt.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:52 +0000 UTC Normal Pod check-span-7vtkt.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 251ms (251ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:52 +0000 UTC Normal Pod check-span-7vtkt.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:52 +0000 UTC Normal Pod check-span-7vtkt.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:52 +0000 UTC Normal Pod report-span-rbmzv AddedInterface Add eth0 [10.129.2.82/23] from ovn-kubernetes multus logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:52 +0000 UTC Normal Pod report-span-rbmzv.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:52 +0000 UTC Normal Pod report-span-rbmzv.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 196ms (196ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:52 +0000 UTC Normal Pod report-span-rbmzv.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:30:52 +0000 UTC Normal Pod report-span-rbmzv.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 18:31:04 | examples-all-in-one-with-options | 2024-09-30 18:31:03 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 18:31:04 | examples-all-in-one-with-options | Deleting namespace: kuttl-test-apt-mallard === CONT kuttl/harness/examples-agent-with-priority-class logger.go:42: 18:31:16 | examples-agent-with-priority-class | Creating namespace: kuttl-test-keen-quetzal logger.go:42: 18:31:16 | examples-agent-with-priority-class/0-install | starting test step 0-install logger.go:42: 18:31:17 | examples-agent-with-priority-class/0-install | SecurityContextConstraints:/daemonset-with-hostport created logger.go:42: 18:31:17 | examples-agent-with-priority-class/0-install | ServiceAccount:kuttl-test-keen-quetzal/jaeger-agent-daemonset created logger.go:42: 18:31:17 | examples-agent-with-priority-class/0-install | test step completed 0-install logger.go:42: 18:31:17 | examples-agent-with-priority-class/1-add-policy | starting test step 1-add-policy logger.go:42: 18:31:17 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c oc adm policy --namespace $NAMESPACE add-scc-to-user daemonset-with-hostport -z jaeger-agent-daemonset] logger.go:42: 18:31:17 | examples-agent-with-priority-class/1-add-policy | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:daemonset-with-hostport added: "jaeger-agent-daemonset" logger.go:42: 18:31:17 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c sleep 5] logger.go:42: 18:31:22 | examples-agent-with-priority-class/1-add-policy | test step completed 1-add-policy logger.go:42: 18:31:22 | examples-agent-with-priority-class/2-install | starting test step 2-install logger.go:42: 18:31:22 | examples-agent-with-priority-class/2-install | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE agent-as-daemonset /dev/null] logger.go:42: 18:31:30 | examples-agent-with-priority-class/2-install | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 18:31:30 | examples-agent-with-priority-class/2-install | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 18:31:31 | examples-agent-with-priority-class/2-install | job.batch/report-span created logger.go:42: 18:31:31 | examples-agent-with-priority-class/2-install | job.batch/check-span created logger.go:42: 18:31:31 | examples-agent-with-priority-class/2-install | PriorityClass:/high-priority created logger.go:42: 18:31:31 | examples-agent-with-priority-class/2-install | Jaeger:kuttl-test-keen-quetzal/agent-as-daemonset updated logger.go:42: 18:31:43 | examples-agent-with-priority-class/2-install | test step completed 2-install logger.go:42: 18:31:43 | examples-agent-with-priority-class | examples-agent-with-priority-class events from ns kuttl-test-keen-quetzal: logger.go:42: 18:31:43 | examples-agent-with-priority-class | 2024-09-30 18:31:27 +0000 UTC Normal Pod agent-as-daemonset-576cdb7867-nbgtv Binding Scheduled Successfully assigned kuttl-test-keen-quetzal/agent-as-daemonset-576cdb7867-nbgtv to ip-10-0-64-30.ec2.internal default-scheduler logger.go:42: 18:31:43 | examples-agent-with-priority-class | 2024-09-30 18:31:27 +0000 UTC Normal ReplicaSet.apps agent-as-daemonset-576cdb7867 SuccessfulCreate Created pod: agent-as-daemonset-576cdb7867-nbgtv replicaset-controller logger.go:42: 18:31:43 | examples-agent-with-priority-class | 2024-09-30 18:31:27 +0000 UTC Normal Deployment.apps agent-as-daemonset ScalingReplicaSet Scaled up replica set agent-as-daemonset-576cdb7867 to 1 deployment-controller logger.go:42: 18:31:43 | examples-agent-with-priority-class | 2024-09-30 18:31:28 +0000 UTC Normal Pod agent-as-daemonset-576cdb7867-nbgtv AddedInterface Add eth0 [10.131.0.92/23] from ovn-kubernetes multus logger.go:42: 18:31:43 | examples-agent-with-priority-class | 2024-09-30 18:31:28 +0000 UTC Normal Pod agent-as-daemonset-576cdb7867-nbgtv.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 18:31:43 | examples-agent-with-priority-class | 2024-09-30 18:31:28 +0000 UTC Normal Pod agent-as-daemonset-576cdb7867-nbgtv.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 18:31:43 | examples-agent-with-priority-class | 2024-09-30 18:31:28 +0000 UTC Normal Pod agent-as-daemonset-576cdb7867-nbgtv.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 18:31:43 | examples-agent-with-priority-class | 2024-09-30 18:31:28 +0000 UTC Normal Pod agent-as-daemonset-576cdb7867-nbgtv.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 18:31:43 | examples-agent-with-priority-class | 2024-09-30 18:31:28 +0000 UTC Normal Pod agent-as-daemonset-576cdb7867-nbgtv.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 18:31:43 | examples-agent-with-priority-class | 2024-09-30 18:31:28 +0000 UTC Normal Pod agent-as-daemonset-576cdb7867-nbgtv.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 18:31:43 | examples-agent-with-priority-class | 2024-09-30 18:31:31 +0000 UTC Normal Pod check-span-xc7ft Binding Scheduled Successfully assigned kuttl-test-keen-quetzal/check-span-xc7ft to ip-10-0-14-134.ec2.internal default-scheduler logger.go:42: 18:31:43 | examples-agent-with-priority-class | 2024-09-30 18:31:31 +0000 UTC Normal Pod check-span-xc7ft AddedInterface Add eth0 [10.128.2.50/23] from ovn-kubernetes multus logger.go:42: 18:31:43 | examples-agent-with-priority-class | 2024-09-30 18:31:31 +0000 UTC Normal Pod check-span-xc7ft.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:31:43 | examples-agent-with-priority-class | 2024-09-30 18:31:31 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-xc7ft job-controller logger.go:42: 18:31:43 | examples-agent-with-priority-class | 2024-09-30 18:31:31 +0000 UTC Normal Pod report-span-tsnts Binding Scheduled Successfully assigned kuttl-test-keen-quetzal/report-span-tsnts to ip-10-0-12-212.ec2.internal default-scheduler logger.go:42: 18:31:43 | examples-agent-with-priority-class | 2024-09-30 18:31:31 +0000 UTC Normal Pod report-span-tsnts AddedInterface Add eth0 [10.129.2.83/23] from ovn-kubernetes multus logger.go:42: 18:31:43 | examples-agent-with-priority-class | 2024-09-30 18:31:31 +0000 UTC Normal Pod report-span-tsnts.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 18:31:43 | examples-agent-with-priority-class | 2024-09-30 18:31:31 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-tsnts job-controller logger.go:42: 18:31:43 | examples-agent-with-priority-class | 2024-09-30 18:31:32 +0000 UTC Normal Pod check-span-xc7ft.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 296ms (297ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:31:43 | examples-agent-with-priority-class | 2024-09-30 18:31:32 +0000 UTC Normal Pod check-span-xc7ft.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 18:31:43 | examples-agent-with-priority-class | 2024-09-30 18:31:32 +0000 UTC Normal Pod check-span-xc7ft.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 18:31:43 | examples-agent-with-priority-class | 2024-09-30 18:31:32 +0000 UTC Normal Pod report-span-tsnts.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 341ms (341ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 18:31:43 | examples-agent-with-priority-class | 2024-09-30 18:31:32 +0000 UTC Normal Pod report-span-tsnts.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 18:31:43 | examples-agent-with-priority-class | 2024-09-30 18:31:32 +0000 UTC Normal Pod report-span-tsnts.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 18:31:43 | examples-agent-with-priority-class | 2024-09-30 18:31:35 +0000 UTC Warning DaemonSet.apps agent-as-daemonset-agent-daemonset FailedCreate Error creating: pods "agent-as-daemonset-agent-daemonset-" is forbidden: unable to validate against any security context constraint: [provider "anyuid": Forbidden: not usable by user or serviceaccount, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 5775: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 5778: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 6831: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 6832: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 14271: Host ports are not allowed to be used, provider "restricted": Forbidden: not usable by user or serviceaccount, provider "nonroot-v2": Forbidden: not usable by user or serviceaccount, provider "nonroot": Forbidden: not usable by user or serviceaccount, provider "hostmount-anyuid": Forbidden: not usable by user or serviceaccount, provider "elasticsearch-scc": Forbidden: not usable by user or serviceaccount, provider "machine-api-termination-handler": Forbidden: not usable by user or serviceaccount, provider "daemonset-with-hostport": Forbidden: not usable by user or serviceaccount, provider "hostnetwork-v2": Forbidden: not usable by user or serviceaccount, provider "hostnetwork": Forbidden: not usable by user or serviceaccount, provider "hostaccess": Forbidden: not usable by user or serviceaccount, provider "node-exporter": Forbidden: not usable by user or serviceaccount, provider "privileged": Forbidden: not usable by user or serviceaccount] daemonset-controller logger.go:42: 18:31:43 | examples-agent-with-priority-class | 2024-09-30 18:31:43 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 18:31:43 | examples-agent-with-priority-class | Deleting namespace: kuttl-test-keen-quetzal === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (529.27s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.10s) --- PASS: kuttl/harness/examples-simple-prod (73.07s) --- PASS: kuttl/harness/examples-with-sampling (56.89s) --- PASS: kuttl/harness/examples-with-cassandra (54.72s) --- PASS: kuttl/harness/examples-with-badger (40.49s) --- PASS: kuttl/harness/examples-simplest (39.19s) --- PASS: kuttl/harness/examples-simple-prod-with-volumes (73.50s) --- PASS: kuttl/harness/examples-business-application-injected-sidecar (45.16s) --- PASS: kuttl/harness/examples-service-types (40.96s) --- PASS: kuttl/harness/examples-openshift-with-htpasswd (24.89s) --- PASS: kuttl/harness/examples-all-in-one-with-options (40.31s) --- PASS: kuttl/harness/examples-agent-with-priority-class (33.83s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name examples --report --output /logs/artifacts/examples.xml ./artifacts/kuttl-report.xml time="2024-09-30T18:31:51Z" level=debug msg="Setting a new name for the test suites" time="2024-09-30T18:31:51Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-09-30T18:31:51Z" level=debug msg="normalizing test case names" time="2024-09-30T18:31:51Z" level=debug msg="examples/artifacts -> examples_artifacts" time="2024-09-30T18:31:51Z" level=debug msg="examples/examples-simple-prod -> examples_examples_simple_prod" time="2024-09-30T18:31:51Z" level=debug msg="examples/examples-with-sampling -> examples_examples_with_sampling" time="2024-09-30T18:31:51Z" level=debug msg="examples/examples-with-cassandra -> examples_examples_with_cassandra" time="2024-09-30T18:31:51Z" level=debug msg="examples/examples-with-badger -> examples_examples_with_badger" time="2024-09-30T18:31:51Z" level=debug msg="examples/examples-simplest -> examples_examples_simplest" time="2024-09-30T18:31:51Z" level=debug msg="examples/examples-simple-prod-with-volumes -> examples_examples_simple_prod_with_volumes" time="2024-09-30T18:31:51Z" level=debug msg="examples/examples-business-application-injected-sidecar -> examples_examples_business_application_injected_sidecar" time="2024-09-30T18:31:51Z" level=debug msg="examples/examples-service-types -> examples_examples_service_types" time="2024-09-30T18:31:51Z" level=debug msg="examples/examples-openshift-with-htpasswd -> examples_examples_openshift_with_htpasswd" time="2024-09-30T18:31:51Z" level=debug msg="examples/examples-all-in-one-with-options -> examples_examples_all_in_one_with_options" time="2024-09-30T18:31:51Z" level=debug msg="examples/examples-agent-with-priority-class -> examples_examples_agent_with_priority_class" +---------------------------------------------------------+--------+ | NAME | RESULT | +---------------------------------------------------------+--------+ | examples_artifacts | passed | | examples_examples_simple_prod | passed | | examples_examples_with_sampling | passed | | examples_examples_with_cassandra | passed | | examples_examples_with_badger | passed | | examples_examples_simplest | passed | | examples_examples_simple_prod_with_volumes | passed | | examples_examples_business_application_injected_sidecar | passed | | examples_examples_service_types | passed | | examples_examples_openshift_with_htpasswd | passed | | examples_examples_all_in_one_with_options | passed | | examples_examples_agent_with_priority_class | passed | +---------------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh generate false true + '[' 3 -ne 3 ']' + test_suite_name=generate + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/generate.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-generate make[2]: Entering directory '/tmp/jaeger-tests' test -s /tmp/jaeger-tests/bin/operator-sdk || curl -sLo /tmp/jaeger-tests/bin/operator-sdk https://github.com/operator-framework/operator-sdk/releases/download/v1.32.0/operator-sdk_`go env GOOS`_`go env GOARCH` ./hack/install/install-golangci-lint.sh Installing golangci-lint golangci-lint 1.55.2 is installed already ./hack/install/install-goimports.sh Installing goimports Try 0... go install golang.org/x/tools/cmd/goimports@v0.1.12 >>>> Formatting code... ./.ci/format.sh >>>> Building... ./hack/install/install-dependencies.sh Installing go dependencies Try 0... go mod download GOOS= GOARCH= CGO_ENABLED=0 GO111MODULE=on go build -ldflags "-X "github.com/jaegertracing/jaeger-operator/pkg/version".version="1.61.0" -X "github.com/jaegertracing/jaeger-operator/pkg/version".buildDate=2024-09-30T18:31:52Z -X "github.com/jaegertracing/jaeger-operator/pkg/version".defaultJaeger="1.61.0"" -o "bin/jaeger-operator" main.go JAEGER_VERSION="1.61.0" ./tests/e2e/generate/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.17.0-0.nightly-2024-09-29-173421 True False 54m Cluster version is 4.17.0-0.nightly-2024-09-29-173421' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.17.0-0.nightly-2024-09-29-173421 True False 54m Cluster version is 4.17.0-0.nightly-2024-09-29-173421' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/generate/render.sh ++ export SUITE_DIR=./tests/e2e/generate ++ SUITE_DIR=./tests/e2e/generate ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/generate ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + '[' true = true ']' + skip_test generate 'This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 2 -ne 2 ']' + test_name=generate + message='This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/generate/_build + '[' _build '!=' _build ']' + rm -rf generate + warning 'generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed\e[0m' WAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running generate E2E tests' Running generate E2E tests + cd tests/e2e/generate/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2856073267 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 18:32:12 | artifacts | Creating namespace: kuttl-test-flexible-whale logger.go:42: 18:32:13 | artifacts | artifacts events from ns kuttl-test-flexible-whale: logger.go:42: 18:32:13 | artifacts | Deleting namespace: kuttl-test-flexible-whale === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (6.15s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.98s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name generate --report --output /logs/artifacts/generate.xml ./artifacts/kuttl-report.xml time="2024-09-30T18:32:19Z" level=debug msg="Setting a new name for the test suites" time="2024-09-30T18:32:19Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-09-30T18:32:19Z" level=debug msg="normalizing test case names" time="2024-09-30T18:32:19Z" level=debug msg="generate/artifacts -> generate_artifacts" +--------------------+--------+ | NAME | RESULT | +--------------------+--------+ | generate_artifacts | passed | +--------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh upgrade false true + '[' 3 -ne 3 ']' + test_suite_name=upgrade + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/upgrade.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-upgrade make[2]: Entering directory '/tmp/jaeger-tests' make docker JAEGER_VERSION=1.61.1 IMG="quay.io//jaeger-operator:next" make[3]: Entering directory '/tmp/jaeger-tests' [ ! -z "true" ] || docker build --build-arg=GOPROXY= --build-arg=VERSION="1.61.0" --build-arg=JAEGER_VERSION=1.61.1 --build-arg=TARGETARCH= --build-arg VERSION_DATE=2024-09-30T18:32:19Z --build-arg VERSION_PKG="github.com/jaegertracing/jaeger-operator/pkg/version" -t "quay.io//jaeger-operator:next" . make[3]: Leaving directory '/tmp/jaeger-tests' touch build-e2e-upgrade-image SKIP_ES_EXTERNAL=true IMG=quay.io//jaeger-operator:"1.61.0" JAEGER_OPERATOR_VERSION="1.61.0" JAEGER_VERSION="1.61.0" ./tests/e2e/upgrade/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.17.0-0.nightly-2024-09-29-173421 True False 55m Cluster version is 4.17.0-0.nightly-2024-09-29-173421' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.17.0-0.nightly-2024-09-29-173421 True False 55m Cluster version is 4.17.0-0.nightly-2024-09-29-173421' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/upgrade/render.sh ++ export SUITE_DIR=./tests/e2e/upgrade ++ SUITE_DIR=./tests/e2e/upgrade ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/upgrade ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + export JAEGER_NAME + '[' true = true ']' + skip_test upgrade 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade + warning 'upgrade: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade: Test not supported in OpenShift\e[0m' WAR: upgrade: Test not supported in OpenShift + '[' true = true ']' + skip_test upgrade-from-latest-release 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade-from-latest-release + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade-from-latest-release + warning 'upgrade-from-latest-release: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade-from-latest-release: Test not supported in OpenShift\e[0m' WAR: upgrade-from-latest-release: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running upgrade E2E tests' Running upgrade E2E tests + cd tests/e2e/upgrade/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2856073267 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-0j41p2wz-016d3.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 18:32:20 | artifacts | Creating namespace: kuttl-test-willing-sparrow logger.go:42: 18:32:20 | artifacts | artifacts events from ns kuttl-test-willing-sparrow: logger.go:42: 18:32:20 | artifacts | Deleting namespace: kuttl-test-willing-sparrow === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (6.06s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.89s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name upgrade --report --output /logs/artifacts/upgrade.xml ./artifacts/kuttl-report.xml time="2024-09-30T18:32:26Z" level=debug msg="Setting a new name for the test suites" time="2024-09-30T18:32:26Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-09-30T18:32:26Z" level=debug msg="normalizing test case names" time="2024-09-30T18:32:26Z" level=debug msg="upgrade/artifacts -> upgrade_artifacts" +-------------------+--------+ | NAME | RESULT | +-------------------+--------+ | upgrade_artifacts | passed | +-------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests'