% Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 3831 100 3831 0 0 24116 0 --:--:-- --:--:-- --:--:-- 24246 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 9315 100 9315 0 0 67821 0 --:--:-- --:--:-- --:--:-- 67992 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 2607 100 2607 0 0 25835 0 --:--:-- --:--:-- --:--:-- 26070 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 1953 100 1953 0 0 20933 0 --:--:-- --:--:-- --:--:-- 21000 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 350 100 350 0 0 3764 0 --:--:-- --:--:-- --:--:-- 3763 100 350 100 350 0 0 3762 0 --:--:-- --:--:-- --:--:-- 3763 Installing kuttl Try 0... curl -sLo /tmp/jaeger-tests/hack/install/../../bin/kubectl-kuttl https://github.com/kudobuilder/kuttl/releases/download/v0.15.0/kubectl-kuttl_0.15.0_linux_x86_64 KUBECONFIG file is: /tmp/kubeconfig-559878434 for suite in miscellaneous elasticsearch examples generate upgrade sidecar streaming ui; do \ make run-e2e-tests-$suite ; \ done make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh miscellaneous false true + '[' 3 -ne 3 ']' + test_suite_name=miscellaneous + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/miscellaneous.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-miscellaneous make[2]: Entering directory '/tmp/jaeger-tests' SKIP_ES_EXTERNAL=true ./tests/e2e/miscellaneous/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-05-19-083311 True False 8m9s Cluster version is 4.16.0-0.nightly-2024-05-19-083311' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-05-19-083311 True False 8m9s Cluster version is 4.16.0-0.nightly-2024-05-19-083311' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/miscellaneous/render.sh ++ export SUITE_DIR=./tests/e2e/miscellaneous ++ SUITE_DIR=./tests/e2e/miscellaneous ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/miscellaneous ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test collector-autoscale + '[' 1 -ne 1 ']' + test_name=collector-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-autoscale\e[0m' Rendering files for test collector-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p collector-autoscale + cd collector-autoscale + jaeger_name=simple-prod + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + ELASTICSEARCH_NODECOUNT=1 + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.resources.requests.memory="200m"' 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.autoscale=true 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.minReplicas=1 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.maxReplicas=2 01-install.yaml + version_lt 1.27 1.23 ++ echo 1.27 1.23 ++ tr ' ' '\n' ++ sort -rV ++ head -n 1 + test 1.27 '!=' 1.27 + rm ./03-assert.yaml + generate_otlp_e2e_tests http + test_protocol=http + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-http\e[0m' Rendering files for test collector-otlp-allinone-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-autoscale + '[' collector-autoscale '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-http + cd collector-otlp-allinone-http + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger http true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-http\e[0m' Rendering files for test collector-otlp-production-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-http + '[' collector-otlp-allinone-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-http + cd collector-otlp-production-http + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger http true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + generate_otlp_e2e_tests grpc + test_protocol=grpc + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-grpc\e[0m' Rendering files for test collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-http + '[' collector-otlp-production-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-grpc + cd collector-otlp-allinone-grpc + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-grpc\e[0m' Rendering files for test collector-otlp-production-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-grpc + '[' collector-otlp-allinone-grpc '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-grpc + cd collector-otlp-production-grpc + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + '[' true = true ']' + skip_test istio 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=istio + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-grpc + '[' collector-otlp-production-grpc '!=' _build ']' + cd .. + rm -rf istio + warning 'istio: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: istio: Test not supported in OpenShift\e[0m' WAR: istio: Test not supported in OpenShift + '[' true = true ']' + skip_test outside-cluster 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=outside-cluster + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + rm -rf outside-cluster + warning 'outside-cluster: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: outside-cluster: Test not supported in OpenShift\e[0m' WAR: outside-cluster: Test not supported in OpenShift + start_test set-custom-img + '[' 1 -ne 1 ']' + test_name=set-custom-img + echo =========================================================================== =========================================================================== + info 'Rendering files for test set-custom-img' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test set-custom-img\e[0m' Rendering files for test set-custom-img + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p set-custom-img + cd set-custom-img + jaeger_name=my-jaeger + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.image="test"' ./02-install.yaml + '[' true = true ']' + skip_test non-cluster-wide 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=non-cluster-wide + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/set-custom-img + '[' set-custom-img '!=' _build ']' + cd .. + rm -rf non-cluster-wide + warning 'non-cluster-wide: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: non-cluster-wide: Test not supported in OpenShift\e[0m' WAR: non-cluster-wide: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running miscellaneous E2E tests' Running miscellaneous E2E tests + cd tests/e2e/miscellaneous/_build + set +e + KUBECONFIG=/tmp/kubeconfig-559878434 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 8 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/cassandra-spark === PAUSE kuttl/harness/cassandra-spark === RUN kuttl/harness/collector-autoscale === PAUSE kuttl/harness/collector-autoscale === RUN kuttl/harness/collector-otlp-allinone-grpc === PAUSE kuttl/harness/collector-otlp-allinone-grpc === RUN kuttl/harness/collector-otlp-allinone-http === PAUSE kuttl/harness/collector-otlp-allinone-http === RUN kuttl/harness/collector-otlp-production-grpc === PAUSE kuttl/harness/collector-otlp-production-grpc === RUN kuttl/harness/collector-otlp-production-http === PAUSE kuttl/harness/collector-otlp-production-http === RUN kuttl/harness/set-custom-img === PAUSE kuttl/harness/set-custom-img === CONT kuttl/harness/artifacts logger.go:42: 07:50:14 | artifacts | Creating namespace: kuttl-test-enhanced-shrimp logger.go:42: 07:50:14 | artifacts | artifacts events from ns kuttl-test-enhanced-shrimp: logger.go:42: 07:50:14 | artifacts | Deleting namespace: kuttl-test-enhanced-shrimp === CONT kuttl/harness/collector-otlp-allinone-http logger.go:42: 07:50:20 | collector-otlp-allinone-http | Creating namespace: kuttl-test-dynamic-donkey logger.go:42: 07:50:20 | collector-otlp-allinone-http/0-install | starting test step 0-install logger.go:42: 07:50:20 | collector-otlp-allinone-http/0-install | Jaeger:kuttl-test-dynamic-donkey/my-jaeger created logger.go:42: 07:50:35 | collector-otlp-allinone-http/0-install | test step completed 0-install logger.go:42: 07:50:35 | collector-otlp-allinone-http/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:50:35 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:50:37 | collector-otlp-allinone-http/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:50:45 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337 OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:50:46 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:50:46 | collector-otlp-allinone-http/1-smoke-test | job.batch/report-span created logger.go:42: 07:50:46 | collector-otlp-allinone-http/1-smoke-test | job.batch/check-span created logger.go:42: 07:51:00 | collector-otlp-allinone-http/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:51:00 | collector-otlp-allinone-http | collector-otlp-allinone-http events from ns kuttl-test-dynamic-donkey: logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:24 +0000 UTC Normal Pod my-jaeger-77d46644cc-7cv8b Binding Scheduled Successfully assigned kuttl-test-dynamic-donkey/my-jaeger-77d46644cc-7cv8b to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:24 +0000 UTC Normal ReplicaSet.apps my-jaeger-77d46644cc SuccessfulCreate Created pod: my-jaeger-77d46644cc-7cv8b replicaset-controller logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:24 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-77d46644cc to 1 deployment-controller logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:25 +0000 UTC Normal Pod my-jaeger-77d46644cc-7cv8b AddedInterface Add eth0 [10.129.2.13/23] from ovn-kubernetes multus logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:25 +0000 UTC Normal Pod my-jaeger-77d46644cc-7cv8b.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" kubelet logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:28 +0000 UTC Normal Pod my-jaeger-77d46644cc-7cv8b.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" in 3.39s (3.39s including waiting) kubelet logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:28 +0000 UTC Normal Pod my-jaeger-77d46644cc-7cv8b.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:28 +0000 UTC Normal Pod my-jaeger-77d46644cc-7cv8b.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:28 +0000 UTC Normal Pod my-jaeger-77d46644cc-7cv8b.spec.containers{oauth-proxy} Pulling Pulling image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" kubelet logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:34 +0000 UTC Normal Pod my-jaeger-77d46644cc-7cv8b.spec.containers{oauth-proxy} Pulled Successfully pulled image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" in 5.481s (5.481s including waiting) kubelet logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:34 +0000 UTC Normal Pod my-jaeger-77d46644cc-7cv8b.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:34 +0000 UTC Normal Pod my-jaeger-77d46644cc-7cv8b.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:39 +0000 UTC Normal Pod my-jaeger-5776b8b5bb-p2phj Binding Scheduled Successfully assigned kuttl-test-dynamic-donkey/my-jaeger-5776b8b5bb-p2phj to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:39 +0000 UTC Normal ReplicaSet.apps my-jaeger-5776b8b5bb SuccessfulCreate Created pod: my-jaeger-5776b8b5bb-p2phj replicaset-controller logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:39 +0000 UTC Normal Pod my-jaeger-77d46644cc-7cv8b.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:39 +0000 UTC Normal Pod my-jaeger-77d46644cc-7cv8b.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:39 +0000 UTC Normal ReplicaSet.apps my-jaeger-77d46644cc SuccessfulDelete Deleted pod: my-jaeger-77d46644cc-7cv8b replicaset-controller logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:39 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-77d46644cc to 0 from 1 deployment-controller logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:39 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-5776b8b5bb to 1 deployment-controller logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:40 +0000 UTC Normal Pod my-jaeger-5776b8b5bb-p2phj AddedInterface Add eth0 [10.129.2.14/23] from ovn-kubernetes multus logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:40 +0000 UTC Normal Pod my-jaeger-5776b8b5bb-p2phj.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" already present on machine kubelet logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:40 +0000 UTC Normal Pod my-jaeger-5776b8b5bb-p2phj.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:40 +0000 UTC Normal Pod my-jaeger-5776b8b5bb-p2phj.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:40 +0000 UTC Normal Pod my-jaeger-5776b8b5bb-p2phj.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:40 +0000 UTC Normal Pod my-jaeger-5776b8b5bb-p2phj.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:40 +0000 UTC Normal Pod my-jaeger-5776b8b5bb-p2phj.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:46 +0000 UTC Normal Pod check-span-zpnrj Binding Scheduled Successfully assigned kuttl-test-dynamic-donkey/check-span-zpnrj to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:46 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-zpnrj job-controller logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:46 +0000 UTC Normal Pod report-span-w9rgs Binding Scheduled Successfully assigned kuttl-test-dynamic-donkey/report-span-w9rgs to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:46 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-w9rgs job-controller logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:47 +0000 UTC Normal Pod check-span-zpnrj AddedInterface Add eth0 [10.128.2.28/23] from ovn-kubernetes multus logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:47 +0000 UTC Normal Pod check-span-zpnrj.spec.containers{asserts-container} Pulling Pulling image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" kubelet logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:47 +0000 UTC Normal Pod report-span-w9rgs AddedInterface Add eth0 [10.131.0.19/23] from ovn-kubernetes multus logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:47 +0000 UTC Normal Pod report-span-w9rgs.spec.containers{report-span} Pulling Pulling image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" kubelet logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:49 +0000 UTC Normal Pod check-span-zpnrj.spec.containers{asserts-container} Pulled Successfully pulled image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" in 2.119s (2.119s including waiting) kubelet logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:49 +0000 UTC Normal Pod check-span-zpnrj.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:49 +0000 UTC Normal Pod check-span-zpnrj.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:49 +0000 UTC Normal Pod report-span-w9rgs.spec.containers{report-span} Pulled Successfully pulled image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" in 2.429s (2.429s including waiting) kubelet logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:49 +0000 UTC Normal Pod report-span-w9rgs.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:50:49 +0000 UTC Normal Pod report-span-w9rgs.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:51:00 | collector-otlp-allinone-http | 2024-05-20 07:51:00 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:51:00 | collector-otlp-allinone-http | Deleting namespace: kuttl-test-dynamic-donkey === CONT kuttl/harness/set-custom-img logger.go:42: 07:51:07 | set-custom-img | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:51:07 | set-custom-img | Ignoring check-collector-img.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:51:07 | set-custom-img | Creating namespace: kuttl-test-safe-camel logger.go:42: 07:51:07 | set-custom-img/1-install | starting test step 1-install logger.go:42: 07:51:07 | set-custom-img/1-install | Jaeger:kuttl-test-safe-camel/my-jaeger created logger.go:42: 07:52:10 | set-custom-img/1-install | test step completed 1-install logger.go:42: 07:52:10 | set-custom-img/2-install | starting test step 2-install logger.go:42: 07:52:10 | set-custom-img/2-install | Jaeger:kuttl-test-safe-camel/my-jaeger updated logger.go:42: 07:52:10 | set-custom-img/2-install | test step completed 2-install logger.go:42: 07:52:10 | set-custom-img/3-check-image | starting test step 3-check-image logger.go:42: 07:52:10 | set-custom-img/3-check-image | running command: [sh -c ./check-collector-img.sh] logger.go:42: 07:52:10 | set-custom-img/3-check-image | Collector image missmatch. Expected: test. Has: registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:261a2c64c771323ad259e14802703bb790ee6b486b72c4939b487fbd83c30703 logger.go:42: 07:52:15 | set-custom-img/3-check-image | Collector image asserted properly! logger.go:42: 07:52:15 | set-custom-img/3-check-image | test step completed 3-check-image logger.go:42: 07:52:15 | set-custom-img | set-custom-img events from ns kuttl-test-safe-camel: logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:51:15 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsafecamelmyjaeger-1-67f9cf446d-7zdcf Binding Scheduled Successfully assigned kuttl-test-safe-camel/elasticsearch-cdm-kuttltestsafecamelmyjaeger-1-67f9cf446d-7zdcf to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:51:15 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestsafecamelmyjaeger-1-67f9cf446d SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestsafecamelmyjaeger-1-67f9cf446d-7zdcf replicaset-controller logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:51:15 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestsafecamelmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestsafecamelmyjaeger-1-67f9cf446d to 1 deployment-controller logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:51:16 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsafecamelmyjaeger-1-67f9cf446d-7zdcf AddedInterface Add eth0 [10.129.2.15/23] from ovn-kubernetes multus logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:51:16 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsafecamelmyjaeger-1-67f9cf446d-7zdcf.spec.containers{elasticsearch} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" kubelet logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:51:23 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsafecamelmyjaeger-1-67f9cf446d-7zdcf.spec.containers{elasticsearch} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" in 7.148s (7.148s including waiting) kubelet logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:51:23 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsafecamelmyjaeger-1-67f9cf446d-7zdcf.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:51:23 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsafecamelmyjaeger-1-67f9cf446d-7zdcf.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:51:23 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsafecamelmyjaeger-1-67f9cf446d-7zdcf.spec.containers{proxy} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" kubelet logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:51:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsafecamelmyjaeger-1-67f9cf446d-7zdcf.spec.containers{proxy} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" in 2.336s (2.336s including waiting) kubelet logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:51:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsafecamelmyjaeger-1-67f9cf446d-7zdcf.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:51:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsafecamelmyjaeger-1-67f9cf446d-7zdcf.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:51:36 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestsafecamelmyjaeger-1-67f9cf446d-7zdcf.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:51:41 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestsafecamelmyjaeger-1-67f9cf446d-7zdcf.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:51:52 +0000 UTC Normal Pod my-jaeger-collector-75f995f9f6-xxfj9 Binding Scheduled Successfully assigned kuttl-test-safe-camel/my-jaeger-collector-75f995f9f6-xxfj9 to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:51:52 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-75f995f9f6 SuccessfulCreate Created pod: my-jaeger-collector-75f995f9f6-xxfj9 replicaset-controller logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:51:52 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-75f995f9f6 to 1 deployment-controller logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:51:52 +0000 UTC Normal Pod my-jaeger-query-5588668b56-dh8x2 Binding Scheduled Successfully assigned kuttl-test-safe-camel/my-jaeger-query-5588668b56-dh8x2 to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:51:52 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-5588668b56 SuccessfulCreate Created pod: my-jaeger-query-5588668b56-dh8x2 replicaset-controller logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:51:52 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-5588668b56 to 1 deployment-controller logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:51:53 +0000 UTC Normal Pod my-jaeger-collector-75f995f9f6-xxfj9 AddedInterface Add eth0 [10.131.0.20/23] from ovn-kubernetes multus logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:51:53 +0000 UTC Normal Pod my-jaeger-collector-75f995f9f6-xxfj9.spec.containers{jaeger-collector} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:261a2c64c771323ad259e14802703bb790ee6b486b72c4939b487fbd83c30703" kubelet logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:51:53 +0000 UTC Normal Pod my-jaeger-query-5588668b56-dh8x2 AddedInterface Add eth0 [10.128.2.29/23] from ovn-kubernetes multus logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:51:53 +0000 UTC Normal Pod my-jaeger-query-5588668b56-dh8x2.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:156ea63cbd4dcd8c44173f118dba74388adbf1e82493237a4c514bb14131d5b8" kubelet logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:51:57 +0000 UTC Normal Pod my-jaeger-collector-75f995f9f6-xxfj9.spec.containers{jaeger-collector} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:261a2c64c771323ad259e14802703bb790ee6b486b72c4939b487fbd83c30703" in 4.309s (4.309s including waiting) kubelet logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:51:57 +0000 UTC Normal Pod my-jaeger-collector-75f995f9f6-xxfj9.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:51:57 +0000 UTC Normal Pod my-jaeger-collector-75f995f9f6-xxfj9.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:51:58 +0000 UTC Normal Pod my-jaeger-query-5588668b56-dh8x2.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:156ea63cbd4dcd8c44173f118dba74388adbf1e82493237a4c514bb14131d5b8" in 4.843s (4.843s including waiting) kubelet logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:51:58 +0000 UTC Normal Pod my-jaeger-query-5588668b56-dh8x2.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:51:58 +0000 UTC Normal Pod my-jaeger-query-5588668b56-dh8x2.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:51:58 +0000 UTC Normal Pod my-jaeger-query-5588668b56-dh8x2.spec.containers{oauth-proxy} Pulling Pulling image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" kubelet logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:52:04 +0000 UTC Normal Pod my-jaeger-query-5588668b56-dh8x2.spec.containers{oauth-proxy} Pulled Successfully pulled image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" in 5.988s (5.988s including waiting) kubelet logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:52:04 +0000 UTC Normal Pod my-jaeger-query-5588668b56-dh8x2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:52:04 +0000 UTC Normal Pod my-jaeger-query-5588668b56-dh8x2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:52:04 +0000 UTC Normal Pod my-jaeger-query-5588668b56-dh8x2.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" kubelet logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:52:07 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:52:07 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:52:07 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:52:08 +0000 UTC Normal Pod my-jaeger-query-5588668b56-dh8x2.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" in 3.405s (3.405s including waiting) kubelet logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:52:08 +0000 UTC Normal Pod my-jaeger-query-5588668b56-dh8x2.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:52:08 +0000 UTC Normal Pod my-jaeger-query-5588668b56-dh8x2.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:52:11 +0000 UTC Normal Pod my-jaeger-collector-75f995f9f6-xxfj9.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:52:11 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-75f995f9f6 SuccessfulDelete Deleted pod: my-jaeger-collector-75f995f9f6-xxfj9 replicaset-controller logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:52:11 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-75f995f9f6 to 0 from 1 deployment-controller logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:52:12 +0000 UTC Normal Pod my-jaeger-collector-76b6b649b5-v5gpc Binding Scheduled Successfully assigned kuttl-test-safe-camel/my-jaeger-collector-76b6b649b5-v5gpc to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:52:12 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-76b6b649b5 SuccessfulCreate Created pod: my-jaeger-collector-76b6b649b5-v5gpc replicaset-controller logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:52:12 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-76b6b649b5 to 1 deployment-controller logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:52:13 +0000 UTC Normal Pod my-jaeger-collector-76b6b649b5-v5gpc AddedInterface Add eth0 [10.131.0.21/23] from ovn-kubernetes multus logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:52:13 +0000 UTC Normal Pod my-jaeger-collector-76b6b649b5-v5gpc.spec.containers{jaeger-collector} Pulling Pulling image "test" kubelet logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:52:13 +0000 UTC Warning Pod my-jaeger-collector-76b6b649b5-v5gpc.spec.containers{jaeger-collector} Failed Failed to pull image "test": reading manifest latest in docker.io/library/test: requested access to the resource is denied kubelet logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:52:13 +0000 UTC Warning Pod my-jaeger-collector-76b6b649b5-v5gpc.spec.containers{jaeger-collector} Failed Error: ErrImagePull kubelet logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:52:13 +0000 UTC Normal Pod my-jaeger-collector-76b6b649b5-v5gpc.spec.containers{jaeger-collector} BackOff Back-off pulling image "test" kubelet logger.go:42: 07:52:15 | set-custom-img | 2024-05-20 07:52:13 +0000 UTC Warning Pod my-jaeger-collector-76b6b649b5-v5gpc.spec.containers{jaeger-collector} Failed Error: ImagePullBackOff kubelet logger.go:42: 07:52:15 | set-custom-img | Deleting namespace: kuttl-test-safe-camel === CONT kuttl/harness/collector-otlp-production-http logger.go:42: 07:52:22 | collector-otlp-production-http | Creating namespace: kuttl-test-quality-stallion logger.go:42: 07:52:22 | collector-otlp-production-http/1-install | starting test step 1-install logger.go:42: 07:52:22 | collector-otlp-production-http/1-install | Jaeger:kuttl-test-quality-stallion/my-jaeger created logger.go:42: 07:52:58 | collector-otlp-production-http/1-install | test step completed 1-install logger.go:42: 07:52:58 | collector-otlp-production-http/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:52:58 | collector-otlp-production-http/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:52:59 | collector-otlp-production-http/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:53:06 | collector-otlp-production-http/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337 OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:53:06 | collector-otlp-production-http/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:53:07 | collector-otlp-production-http/2-smoke-test | job.batch/report-span created logger.go:42: 07:53:07 | collector-otlp-production-http/2-smoke-test | job.batch/check-span created logger.go:42: 07:53:19 | collector-otlp-production-http/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:53:19 | collector-otlp-production-http | collector-otlp-production-http events from ns kuttl-test-quality-stallion: logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:52:27 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestqualitystallionmyjaeger-1-547f4d6b8 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestqualitystallionmyjaeger-1-547f4fmldk replicaset-controller logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:52:27 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestqualitystallionmyjaeger-1-547f4fmldk Binding Scheduled Successfully assigned kuttl-test-quality-stallion/elasticsearch-cdm-kuttltestqualitystallionmyjaeger-1-547f4fmldk to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:52:27 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestqualitystallionmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestqualitystallionmyjaeger-1-547f4d6b8 to 1 deployment-controller logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:52:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestqualitystallionmyjaeger-1-547f4fmldk AddedInterface Add eth0 [10.129.2.16/23] from ovn-kubernetes multus logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:52:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestqualitystallionmyjaeger-1-547f4fmldk.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:52:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestqualitystallionmyjaeger-1-547f4fmldk.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:52:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestqualitystallionmyjaeger-1-547f4fmldk.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:52:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestqualitystallionmyjaeger-1-547f4fmldk.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:52:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestqualitystallionmyjaeger-1-547f4fmldk.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:52:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestqualitystallionmyjaeger-1-547f4fmldk.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:52:43 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestqualitystallionmyjaeger-1-547f4fmldk.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:52:54 +0000 UTC Normal Pod my-jaeger-collector-5889f94b86-dx2n9 Binding Scheduled Successfully assigned kuttl-test-quality-stallion/my-jaeger-collector-5889f94b86-dx2n9 to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:52:54 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-5889f94b86 SuccessfulCreate Created pod: my-jaeger-collector-5889f94b86-dx2n9 replicaset-controller logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:52:54 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-5889f94b86 to 1 deployment-controller logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:52:54 +0000 UTC Normal Pod my-jaeger-query-b84c55c99-49vlf Binding Scheduled Successfully assigned kuttl-test-quality-stallion/my-jaeger-query-b84c55c99-49vlf to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:52:54 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-b84c55c99 SuccessfulCreate Created pod: my-jaeger-query-b84c55c99-49vlf replicaset-controller logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:52:54 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-b84c55c99 to 1 deployment-controller logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:52:55 +0000 UTC Normal Pod my-jaeger-collector-5889f94b86-dx2n9 AddedInterface Add eth0 [10.131.0.22/23] from ovn-kubernetes multus logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:52:55 +0000 UTC Normal Pod my-jaeger-collector-5889f94b86-dx2n9.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:261a2c64c771323ad259e14802703bb790ee6b486b72c4939b487fbd83c30703" already present on machine kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:52:55 +0000 UTC Normal Pod my-jaeger-collector-5889f94b86-dx2n9.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:52:55 +0000 UTC Normal Pod my-jaeger-collector-5889f94b86-dx2n9.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:52:55 +0000 UTC Normal Pod my-jaeger-query-b84c55c99-49vlf AddedInterface Add eth0 [10.128.2.30/23] from ovn-kubernetes multus logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:52:55 +0000 UTC Normal Pod my-jaeger-query-b84c55c99-49vlf.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:156ea63cbd4dcd8c44173f118dba74388adbf1e82493237a4c514bb14131d5b8" already present on machine kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:52:55 +0000 UTC Normal Pod my-jaeger-query-b84c55c99-49vlf.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:52:55 +0000 UTC Normal Pod my-jaeger-query-b84c55c99-49vlf.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:52:55 +0000 UTC Normal Pod my-jaeger-query-b84c55c99-49vlf.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:52:55 +0000 UTC Normal Pod my-jaeger-query-b84c55c99-49vlf.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:52:55 +0000 UTC Normal Pod my-jaeger-query-b84c55c99-49vlf.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:52:55 +0000 UTC Normal Pod my-jaeger-query-b84c55c99-49vlf.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" already present on machine kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:52:55 +0000 UTC Normal Pod my-jaeger-query-b84c55c99-49vlf.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:52:55 +0000 UTC Normal Pod my-jaeger-query-b84c55c99-49vlf.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:01 +0000 UTC Normal Pod my-jaeger-query-b84c55c99-49vlf.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:01 +0000 UTC Normal Pod my-jaeger-query-b84c55c99-49vlf.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:01 +0000 UTC Normal Pod my-jaeger-query-b84c55c99-49vlf.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:01 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-b84c55c99 SuccessfulDelete Deleted pod: my-jaeger-query-b84c55c99-49vlf replicaset-controller logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:01 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-b84c55c99 to 0 from 1 deployment-controller logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:02 +0000 UTC Normal Pod my-jaeger-query-68d6c5c495-xf2dt Binding Scheduled Successfully assigned kuttl-test-quality-stallion/my-jaeger-query-68d6c5c495-xf2dt to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:02 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-68d6c5c495 SuccessfulCreate Created pod: my-jaeger-query-68d6c5c495-xf2dt replicaset-controller logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:02 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-68d6c5c495 to 1 deployment-controller logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:03 +0000 UTC Normal Pod my-jaeger-query-68d6c5c495-xf2dt AddedInterface Add eth0 [10.128.2.31/23] from ovn-kubernetes multus logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:03 +0000 UTC Normal Pod my-jaeger-query-68d6c5c495-xf2dt.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:156ea63cbd4dcd8c44173f118dba74388adbf1e82493237a4c514bb14131d5b8" already present on machine kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:03 +0000 UTC Normal Pod my-jaeger-query-68d6c5c495-xf2dt.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:03 +0000 UTC Normal Pod my-jaeger-query-68d6c5c495-xf2dt.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:03 +0000 UTC Normal Pod my-jaeger-query-68d6c5c495-xf2dt.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:03 +0000 UTC Normal Pod my-jaeger-query-68d6c5c495-xf2dt.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:03 +0000 UTC Normal Pod my-jaeger-query-68d6c5c495-xf2dt.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:03 +0000 UTC Normal Pod my-jaeger-query-68d6c5c495-xf2dt.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" already present on machine kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:03 +0000 UTC Normal Pod my-jaeger-query-68d6c5c495-xf2dt.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:03 +0000 UTC Normal Pod my-jaeger-query-68d6c5c495-xf2dt.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:07 +0000 UTC Normal Pod check-span-m7jp7 Binding Scheduled Successfully assigned kuttl-test-quality-stallion/check-span-m7jp7 to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:07 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-m7jp7 job-controller logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:07 +0000 UTC Normal Pod report-span-l55h5 Binding Scheduled Successfully assigned kuttl-test-quality-stallion/report-span-l55h5 to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:07 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-l55h5 job-controller logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:08 +0000 UTC Normal Pod check-span-m7jp7 AddedInterface Add eth0 [10.131.0.24/23] from ovn-kubernetes multus logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:08 +0000 UTC Normal Pod check-span-m7jp7.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:08 +0000 UTC Normal Pod check-span-m7jp7.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:08 +0000 UTC Normal Pod check-span-m7jp7.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:08 +0000 UTC Normal Pod report-span-l55h5 AddedInterface Add eth0 [10.131.0.23/23] from ovn-kubernetes multus logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:08 +0000 UTC Normal Pod report-span-l55h5.spec.containers{report-span} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:08 +0000 UTC Normal Pod report-span-l55h5.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:08 +0000 UTC Normal Pod report-span-l55h5.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:09 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:09 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-5889f94b86-dx2n9 horizontal-pod-autoscaler logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:09 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:53:19 | collector-otlp-production-http | 2024-05-20 07:53:19 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:53:19 | collector-otlp-production-http | Deleting namespace: kuttl-test-quality-stallion === CONT kuttl/harness/collector-otlp-production-grpc logger.go:42: 07:53:27 | collector-otlp-production-grpc | Creating namespace: kuttl-test-wired-grouper logger.go:42: 07:53:27 | collector-otlp-production-grpc/1-install | starting test step 1-install logger.go:42: 07:53:27 | collector-otlp-production-grpc/1-install | Jaeger:kuttl-test-wired-grouper/my-jaeger created logger.go:42: 07:54:03 | collector-otlp-production-grpc/1-install | test step completed 1-install logger.go:42: 07:54:03 | collector-otlp-production-grpc/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:54:03 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:54:05 | collector-otlp-production-grpc/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:54:12 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337 OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:54:13 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:54:15 | collector-otlp-production-grpc/2-smoke-test | job.batch/report-span created logger.go:42: 07:54:15 | collector-otlp-production-grpc/2-smoke-test | job.batch/check-span created logger.go:42: 07:54:34 | collector-otlp-production-grpc/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:54:34 | collector-otlp-production-grpc | collector-otlp-production-grpc events from ns kuttl-test-wired-grouper: logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:53:33 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestwiredgroupermyjaeger-1-5967b64758 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestwiredgroupermyjaeger-1-5967b647sxfr6 replicaset-controller logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:53:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwiredgroupermyjaeger-1-5967b647sxfr6 Binding Scheduled Successfully assigned kuttl-test-wired-grouper/elasticsearch-cdm-kuttltestwiredgroupermyjaeger-1-5967b647sxfr6 to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:53:33 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestwiredgroupermyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestwiredgroupermyjaeger-1-5967b64758 to 1 deployment-controller logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:53:34 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwiredgroupermyjaeger-1-5967b647sxfr6 AddedInterface Add eth0 [10.129.2.17/23] from ovn-kubernetes multus logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:53:34 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwiredgroupermyjaeger-1-5967b647sxfr6.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:53:34 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwiredgroupermyjaeger-1-5967b647sxfr6.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:53:34 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwiredgroupermyjaeger-1-5967b647sxfr6.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:53:34 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwiredgroupermyjaeger-1-5967b647sxfr6.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:53:34 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwiredgroupermyjaeger-1-5967b647sxfr6.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:53:34 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwiredgroupermyjaeger-1-5967b647sxfr6.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:53:49 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestwiredgroupermyjaeger-1-5967b647sxfr6.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:00 +0000 UTC Normal Pod my-jaeger-collector-58d6dd58fd-fvspr Binding Scheduled Successfully assigned kuttl-test-wired-grouper/my-jaeger-collector-58d6dd58fd-fvspr to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:00 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-58d6dd58fd SuccessfulCreate Created pod: my-jaeger-collector-58d6dd58fd-fvspr replicaset-controller logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:00 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-58d6dd58fd to 1 deployment-controller logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:00 +0000 UTC Normal Pod my-jaeger-query-fd69bb997-9rp5k Binding Scheduled Successfully assigned kuttl-test-wired-grouper/my-jaeger-query-fd69bb997-9rp5k to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:00 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-fd69bb997 SuccessfulCreate Created pod: my-jaeger-query-fd69bb997-9rp5k replicaset-controller logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:00 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-fd69bb997 to 1 deployment-controller logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:01 +0000 UTC Normal Pod my-jaeger-collector-58d6dd58fd-fvspr AddedInterface Add eth0 [10.131.0.25/23] from ovn-kubernetes multus logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:01 +0000 UTC Normal Pod my-jaeger-collector-58d6dd58fd-fvspr.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:261a2c64c771323ad259e14802703bb790ee6b486b72c4939b487fbd83c30703" already present on machine kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:01 +0000 UTC Normal Pod my-jaeger-collector-58d6dd58fd-fvspr.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:01 +0000 UTC Normal Pod my-jaeger-collector-58d6dd58fd-fvspr.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:01 +0000 UTC Normal Pod my-jaeger-query-fd69bb997-9rp5k AddedInterface Add eth0 [10.128.2.32/23] from ovn-kubernetes multus logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:01 +0000 UTC Normal Pod my-jaeger-query-fd69bb997-9rp5k.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:156ea63cbd4dcd8c44173f118dba74388adbf1e82493237a4c514bb14131d5b8" already present on machine kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:01 +0000 UTC Normal Pod my-jaeger-query-fd69bb997-9rp5k.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:01 +0000 UTC Normal Pod my-jaeger-query-fd69bb997-9rp5k.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:01 +0000 UTC Normal Pod my-jaeger-query-fd69bb997-9rp5k.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:01 +0000 UTC Normal Pod my-jaeger-query-fd69bb997-9rp5k.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:01 +0000 UTC Normal Pod my-jaeger-query-fd69bb997-9rp5k.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:01 +0000 UTC Normal Pod my-jaeger-query-fd69bb997-9rp5k.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" already present on machine kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:01 +0000 UTC Normal Pod my-jaeger-query-fd69bb997-9rp5k.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:01 +0000 UTC Normal Pod my-jaeger-query-fd69bb997-9rp5k.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:09 +0000 UTC Normal Pod my-jaeger-query-fd69bb997-9rp5k.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:09 +0000 UTC Normal Pod my-jaeger-query-fd69bb997-9rp5k.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:09 +0000 UTC Normal Pod my-jaeger-query-fd69bb997-9rp5k.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:09 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-fd69bb997 SuccessfulDelete Deleted pod: my-jaeger-query-fd69bb997-9rp5k replicaset-controller logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:09 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-fd69bb997 to 0 from 1 deployment-controller logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:10 +0000 UTC Normal Pod my-jaeger-query-7df86944bc-kw6z2 Binding Scheduled Successfully assigned kuttl-test-wired-grouper/my-jaeger-query-7df86944bc-kw6z2 to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:10 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-7df86944bc SuccessfulCreate Created pod: my-jaeger-query-7df86944bc-kw6z2 replicaset-controller logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:10 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-7df86944bc to 1 deployment-controller logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:11 +0000 UTC Normal Pod my-jaeger-query-7df86944bc-kw6z2 AddedInterface Add eth0 [10.128.2.33/23] from ovn-kubernetes multus logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:11 +0000 UTC Normal Pod my-jaeger-query-7df86944bc-kw6z2.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:156ea63cbd4dcd8c44173f118dba74388adbf1e82493237a4c514bb14131d5b8" already present on machine kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:11 +0000 UTC Normal Pod my-jaeger-query-7df86944bc-kw6z2.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:11 +0000 UTC Normal Pod my-jaeger-query-7df86944bc-kw6z2.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:11 +0000 UTC Normal Pod my-jaeger-query-7df86944bc-kw6z2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:11 +0000 UTC Normal Pod my-jaeger-query-7df86944bc-kw6z2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:11 +0000 UTC Normal Pod my-jaeger-query-7df86944bc-kw6z2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:11 +0000 UTC Normal Pod my-jaeger-query-7df86944bc-kw6z2.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" already present on machine kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:11 +0000 UTC Normal Pod my-jaeger-query-7df86944bc-kw6z2.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:11 +0000 UTC Normal Pod my-jaeger-query-7df86944bc-kw6z2.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:15 +0000 UTC Normal Pod check-span-qxhxn Binding Scheduled Successfully assigned kuttl-test-wired-grouper/check-span-qxhxn to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:15 +0000 UTC Normal Pod check-span-qxhxn AddedInterface Add eth0 [10.131.0.27/23] from ovn-kubernetes multus logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:15 +0000 UTC Normal Pod check-span-qxhxn.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:15 +0000 UTC Normal Pod check-span-qxhxn.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:15 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-qxhxn job-controller logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:15 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:15 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:15 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:15 +0000 UTC Normal Pod report-span-dwk4r Binding Scheduled Successfully assigned kuttl-test-wired-grouper/report-span-dwk4r to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:15 +0000 UTC Normal Pod report-span-dwk4r AddedInterface Add eth0 [10.131.0.26/23] from ovn-kubernetes multus logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:15 +0000 UTC Normal Pod report-span-dwk4r.spec.containers{report-span} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:15 +0000 UTC Normal Pod report-span-dwk4r.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:15 +0000 UTC Normal Pod report-span-dwk4r.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:15 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-dwk4r job-controller logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:16 +0000 UTC Normal Pod check-span-qxhxn.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:30 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:30 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-58d6dd58fd-fvspr horizontal-pod-autoscaler logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:30 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:54:34 | collector-otlp-production-grpc | 2024-05-20 07:54:34 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:54:34 | collector-otlp-production-grpc | Deleting namespace: kuttl-test-wired-grouper === CONT kuttl/harness/collector-autoscale logger.go:42: 07:54:46 | collector-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:54:46 | collector-autoscale | Creating namespace: kuttl-test-national-ray logger.go:42: 07:54:46 | collector-autoscale/1-install | starting test step 1-install logger.go:42: 07:54:46 | collector-autoscale/1-install | Jaeger:kuttl-test-national-ray/simple-prod created logger.go:42: 07:55:24 | collector-autoscale/1-install | test step completed 1-install logger.go:42: 07:55:24 | collector-autoscale/2- | starting test step 2- logger.go:42: 07:55:24 | collector-autoscale/2- | test step completed 2- logger.go:42: 07:55:24 | collector-autoscale | collector-autoscale events from ns kuttl-test-national-ray: logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:54:53 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestnationalraysimpleprod-1-cdd4d6c8rstn Binding Scheduled Successfully assigned kuttl-test-national-ray/elasticsearch-cdm-kuttltestnationalraysimpleprod-1-cdd4d6c8rstn to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:54:53 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestnationalraysimpleprod-1-cdd4d6cf4 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestnationalraysimpleprod-1-cdd4d6c8rstn replicaset-controller logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:54:53 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestnationalraysimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestnationalraysimpleprod-1-cdd4d6cf4 to 1 deployment-controller logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:54:54 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestnationalraysimpleprod-1-cdd4d6c8rstn FailedMount MountVolume.SetUp failed for volume "elasticsearch-config" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:54:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestnationalraysimpleprod-1-cdd4d6c8rstn AddedInterface Add eth0 [10.129.2.18/23] from ovn-kubernetes multus logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:54:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestnationalraysimpleprod-1-cdd4d6c8rstn.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:54:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestnationalraysimpleprod-1-cdd4d6c8rstn.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:54:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestnationalraysimpleprod-1-cdd4d6c8rstn.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:54:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestnationalraysimpleprod-1-cdd4d6c8rstn.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:54:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestnationalraysimpleprod-1-cdd4d6c8rstn.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:54:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestnationalraysimpleprod-1-cdd4d6c8rstn.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:55:10 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestnationalraysimpleprod-1-cdd4d6c8rstn.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:55:20 +0000 UTC Normal Pod simple-prod-collector-65dffff7d7-bdssk Binding Scheduled Successfully assigned kuttl-test-national-ray/simple-prod-collector-65dffff7d7-bdssk to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:55:20 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-65dffff7d7 SuccessfulCreate Created pod: simple-prod-collector-65dffff7d7-bdssk replicaset-controller logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:55:20 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-65dffff7d7 to 1 deployment-controller logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:55:20 +0000 UTC Normal Pod simple-prod-query-9584bf947-rrjv8 Binding Scheduled Successfully assigned kuttl-test-national-ray/simple-prod-query-9584bf947-rrjv8 to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:55:20 +0000 UTC Normal ReplicaSet.apps simple-prod-query-9584bf947 SuccessfulCreate Created pod: simple-prod-query-9584bf947-rrjv8 replicaset-controller logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:55:20 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-9584bf947 to 1 deployment-controller logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:55:21 +0000 UTC Normal Pod simple-prod-collector-65dffff7d7-bdssk AddedInterface Add eth0 [10.131.0.28/23] from ovn-kubernetes multus logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:55:21 +0000 UTC Normal Pod simple-prod-collector-65dffff7d7-bdssk.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:261a2c64c771323ad259e14802703bb790ee6b486b72c4939b487fbd83c30703" already present on machine kubelet logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:55:21 +0000 UTC Normal Pod simple-prod-collector-65dffff7d7-bdssk.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:55:21 +0000 UTC Normal Pod simple-prod-collector-65dffff7d7-bdssk.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:55:21 +0000 UTC Normal Pod simple-prod-query-9584bf947-rrjv8 AddedInterface Add eth0 [10.128.2.34/23] from ovn-kubernetes multus logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:55:21 +0000 UTC Normal Pod simple-prod-query-9584bf947-rrjv8.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:156ea63cbd4dcd8c44173f118dba74388adbf1e82493237a4c514bb14131d5b8" already present on machine kubelet logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:55:21 +0000 UTC Normal Pod simple-prod-query-9584bf947-rrjv8.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:55:21 +0000 UTC Normal Pod simple-prod-query-9584bf947-rrjv8.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:55:21 +0000 UTC Normal Pod simple-prod-query-9584bf947-rrjv8.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:55:21 +0000 UTC Normal Pod simple-prod-query-9584bf947-rrjv8.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:55:21 +0000 UTC Normal Pod simple-prod-query-9584bf947-rrjv8.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:55:21 +0000 UTC Normal Pod simple-prod-query-9584bf947-rrjv8.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" already present on machine kubelet logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:55:21 +0000 UTC Normal Pod simple-prod-query-9584bf947-rrjv8.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:55:24 | collector-autoscale | 2024-05-20 07:55:22 +0000 UTC Normal Pod simple-prod-query-9584bf947-rrjv8.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:55:24 | collector-autoscale | Deleting namespace: kuttl-test-national-ray === CONT kuttl/harness/collector-otlp-allinone-grpc logger.go:42: 07:55:30 | collector-otlp-allinone-grpc | Creating namespace: kuttl-test-shining-vervet logger.go:42: 07:55:30 | collector-otlp-allinone-grpc/0-install | starting test step 0-install logger.go:42: 07:55:30 | collector-otlp-allinone-grpc/0-install | Jaeger:kuttl-test-shining-vervet/my-jaeger created logger.go:42: 07:55:36 | collector-otlp-allinone-grpc/0-install | test step completed 0-install logger.go:42: 07:55:36 | collector-otlp-allinone-grpc/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:55:36 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:55:38 | collector-otlp-allinone-grpc/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:55:44 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337 OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:55:45 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:55:45 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/report-span created logger.go:42: 07:55:45 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/check-span created logger.go:42: 07:56:05 | collector-otlp-allinone-grpc/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | collector-otlp-allinone-grpc events from ns kuttl-test-shining-vervet: logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:34 +0000 UTC Normal Pod my-jaeger-86644749c6-896g6 Binding Scheduled Successfully assigned kuttl-test-shining-vervet/my-jaeger-86644749c6-896g6 to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:34 +0000 UTC Normal Pod my-jaeger-86644749c6-896g6 AddedInterface Add eth0 [10.129.2.19/23] from ovn-kubernetes multus logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:34 +0000 UTC Normal Pod my-jaeger-86644749c6-896g6.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" already present on machine kubelet logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:34 +0000 UTC Normal ReplicaSet.apps my-jaeger-86644749c6 SuccessfulCreate Created pod: my-jaeger-86644749c6-896g6 replicaset-controller logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:34 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-86644749c6 to 1 deployment-controller logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:35 +0000 UTC Normal Pod my-jaeger-86644749c6-896g6.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:35 +0000 UTC Normal Pod my-jaeger-86644749c6-896g6.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:35 +0000 UTC Normal Pod my-jaeger-86644749c6-896g6.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:35 +0000 UTC Normal Pod my-jaeger-86644749c6-896g6.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:35 +0000 UTC Normal Pod my-jaeger-86644749c6-896g6.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:40 +0000 UTC Normal Pod my-jaeger-86644749c6-896g6.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:40 +0000 UTC Normal Pod my-jaeger-86644749c6-896g6.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:40 +0000 UTC Normal ReplicaSet.apps my-jaeger-86644749c6 SuccessfulDelete Deleted pod: my-jaeger-86644749c6-896g6 replicaset-controller logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:40 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-86644749c6 to 0 from 1 deployment-controller logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:41 +0000 UTC Normal Pod my-jaeger-58b9fc8b9b-94jcg Binding Scheduled Successfully assigned kuttl-test-shining-vervet/my-jaeger-58b9fc8b9b-94jcg to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:41 +0000 UTC Normal ReplicaSet.apps my-jaeger-58b9fc8b9b SuccessfulCreate Created pod: my-jaeger-58b9fc8b9b-94jcg replicaset-controller logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:41 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-58b9fc8b9b to 1 deployment-controller logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:42 +0000 UTC Normal Pod my-jaeger-58b9fc8b9b-94jcg AddedInterface Add eth0 [10.129.2.20/23] from ovn-kubernetes multus logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:42 +0000 UTC Normal Pod my-jaeger-58b9fc8b9b-94jcg.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" already present on machine kubelet logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:42 +0000 UTC Normal Pod my-jaeger-58b9fc8b9b-94jcg.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:42 +0000 UTC Normal Pod my-jaeger-58b9fc8b9b-94jcg.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:42 +0000 UTC Normal Pod my-jaeger-58b9fc8b9b-94jcg.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:42 +0000 UTC Normal Pod my-jaeger-58b9fc8b9b-94jcg.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:42 +0000 UTC Normal Pod my-jaeger-58b9fc8b9b-94jcg.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:45 +0000 UTC Normal Pod check-span-x66pz Binding Scheduled Successfully assigned kuttl-test-shining-vervet/check-span-x66pz to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:45 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-x66pz job-controller logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:45 +0000 UTC Normal Pod report-span-g2cwr Binding Scheduled Successfully assigned kuttl-test-shining-vervet/report-span-g2cwr to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:45 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-g2cwr job-controller logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:46 +0000 UTC Normal Pod check-span-x66pz AddedInterface Add eth0 [10.128.2.35/23] from ovn-kubernetes multus logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:46 +0000 UTC Normal Pod check-span-x66pz.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:46 +0000 UTC Normal Pod check-span-x66pz.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:46 +0000 UTC Normal Pod check-span-x66pz.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:46 +0000 UTC Normal Pod report-span-g2cwr AddedInterface Add eth0 [10.131.0.29/23] from ovn-kubernetes multus logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:46 +0000 UTC Normal Pod report-span-g2cwr.spec.containers{report-span} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:46 +0000 UTC Normal Pod report-span-g2cwr.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:55:46 +0000 UTC Normal Pod report-span-g2cwr.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | 2024-05-20 07:56:04 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:56:05 | collector-otlp-allinone-grpc | Deleting namespace: kuttl-test-shining-vervet === CONT kuttl/harness/cassandra-spark logger.go:42: 07:56:17 | cassandra-spark | Ignoring 01-assert.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:56:17 | cassandra-spark | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:56:17 | cassandra-spark | Creating namespace: kuttl-test-novel-hamster logger.go:42: 07:56:17 | cassandra-spark | cassandra-spark events from ns kuttl-test-novel-hamster: logger.go:42: 07:56:17 | cassandra-spark | Deleting namespace: kuttl-test-novel-hamster === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (369.21s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.93s) --- PASS: kuttl/harness/collector-otlp-allinone-http (46.85s) --- PASS: kuttl/harness/set-custom-img (74.71s) --- PASS: kuttl/harness/collector-otlp-production-http (65.00s) --- PASS: kuttl/harness/collector-otlp-production-grpc (79.72s) --- PASS: kuttl/harness/collector-autoscale (43.58s) --- PASS: kuttl/harness/collector-otlp-allinone-grpc (47.32s) --- PASS: kuttl/harness/cassandra-spark (5.95s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name miscellaneous --report --output /logs/artifacts/miscellaneous.xml ./artifacts/kuttl-report.xml time="2024-05-20T07:56:25Z" level=debug msg="Setting a new name for the test suites" time="2024-05-20T07:56:25Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-05-20T07:56:25Z" level=debug msg="normalizing test case names" time="2024-05-20T07:56:25Z" level=debug msg="miscellaneous/artifacts -> miscellaneous_artifacts" time="2024-05-20T07:56:25Z" level=debug msg="miscellaneous/collector-otlp-allinone-http -> miscellaneous_collector_otlp_allinone_http" time="2024-05-20T07:56:25Z" level=debug msg="miscellaneous/set-custom-img -> miscellaneous_set_custom_img" time="2024-05-20T07:56:25Z" level=debug msg="miscellaneous/collector-otlp-production-http -> miscellaneous_collector_otlp_production_http" time="2024-05-20T07:56:25Z" level=debug msg="miscellaneous/collector-otlp-production-grpc -> miscellaneous_collector_otlp_production_grpc" time="2024-05-20T07:56:25Z" level=debug msg="miscellaneous/collector-autoscale -> miscellaneous_collector_autoscale" time="2024-05-20T07:56:25Z" level=debug msg="miscellaneous/collector-otlp-allinone-grpc -> miscellaneous_collector_otlp_allinone_grpc" time="2024-05-20T07:56:25Z" level=debug msg="miscellaneous/cassandra-spark -> miscellaneous_cassandra_spark" +----------------------------------------------+--------+ | NAME | RESULT | +----------------------------------------------+--------+ | miscellaneous_artifacts | passed | | miscellaneous_collector_otlp_allinone_http | passed | | miscellaneous_set_custom_img | passed | | miscellaneous_collector_otlp_production_http | passed | | miscellaneous_collector_otlp_production_grpc | passed | | miscellaneous_collector_autoscale | passed | | miscellaneous_collector_otlp_allinone_grpc | passed | | miscellaneous_cassandra_spark | passed | +----------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh elasticsearch false true + '[' 3 -ne 3 ']' + test_suite_name=elasticsearch + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/elasticsearch.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-elasticsearch make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true \ KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ ./tests/e2e/elasticsearch/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-05-19-083311 True False 14m Cluster version is 4.16.0-0.nightly-2024-05-19-083311' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-05-19-083311 True False 14m Cluster version is 4.16.0-0.nightly-2024-05-19-083311' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/elasticsearch/render.sh ++ export SUITE_DIR=./tests/e2e/elasticsearch ++ SUITE_DIR=./tests/e2e/elasticsearch ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/elasticsearch ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + start_test es-from-aio-to-production + '[' 1 -ne 1 ']' + test_name=es-from-aio-to-production + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-from-aio-to-production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-from-aio-to-production\e[0m' Rendering files for test es-from-aio-to-production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-from-aio-to-production + cd es-from-aio-to-production + jaeger_name=my-jaeger + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 03 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=03 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./03-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch.redundancyPolicy="ZeroRedundancy"' ./03-install.yaml + render_smoke_test my-jaeger true 04 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=04 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./04-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./04-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test es-increasing-replicas + '[' 1 -ne 1 ']' + test_name=es-increasing-replicas + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-increasing-replicas' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-increasing-replicas\e[0m' Rendering files for test es-increasing-replicas + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-from-aio-to-production + '[' es-from-aio-to-production '!=' _build ']' + cd .. + mkdir -p es-increasing-replicas + cd es-increasing-replicas + jaeger_name=simple-prod + '[' true = true ']' + jaeger_deployment_mode=production_autoprovisioned + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.replicas=2 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.query.replicas=2 ./02-install.yaml + cp ./01-assert.yaml ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.replicas=2 ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .status.readyReplicas=2 ./02-assert.yaml + render_smoke_test simple-prod true 03 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=03 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./03-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + cp ./02-install.yaml ./04-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.elasticsearch.nodeCount=2 ./04-install.yaml + /tmp/jaeger-tests/bin/gomplate -f ./openshift-check-es-nodes.yaml.template -o ./05-check-es-nodes.yaml + '[' true = true ']' + skip_test es-index-cleaner-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-increasing-replicas + '[' es-increasing-replicas '!=' _build ']' + cd .. + rm -rf es-index-cleaner-upstream + warning 'es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_index_cleaner -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-index-cleaner-autoprov + '[' 1 -ne 1 ']' + test_name=es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-index-cleaner-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-index-cleaner-autoprov\e[0m' Rendering files for test es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-index-cleaner-autoprov + cd es-index-cleaner-autoprov + jaeger_name=test-es-index-cleaner-with-prefix + cronjob_name=test-es-index-cleaner-with-prefix-es-index-cleaner + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + cp ../../es-index-cleaner-upstream/04-assert.yaml ../../es-index-cleaner-upstream/README.md . + render_install_jaeger test-es-index-cleaner-with-prefix production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options.es.index-prefix=""' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.enabled=false ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.numberOfDays=0 ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.esIndexCleaner.schedule="*/1 * * * *"' ./01-install.yaml + render_report_spans test-es-index-cleaner-with-prefix true 5 00 true 02 + '[' 6 -ne 6 ']' + jaeger=test-es-index-cleaner-with-prefix + is_secured=true + number_of_spans=5 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + export JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=5 + DAYS=5 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + sed 's~enabled: false~enabled: true~gi' ./01-install.yaml + CRONJOB_NAME=test-es-index-cleaner-with-prefix-es-index-cleaner + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./04-wait-es-index-cleaner.yaml + /tmp/jaeger-tests/bin/gomplate -f ./01-install.yaml -o ./05-install.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 00 06 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=00 + test_step=06 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=test-es-index-cleaner-with-prefix-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=test-es-index-cleaner-with-prefix-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./06-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./06-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.3"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.3"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.3"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.3 ++ version_ge 5.8.3 5.4 +++ echo 5.8.3 5.4 +++ tr ' ' '\n' +++ sort -rV +++ head -n 1 ++ test 5.8.3 == 5.8.3 + '[' -n '' ']' + skip_test es-index-cleaner-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-index-cleaner-autoprov + '[' es-index-cleaner-autoprov '!=' _build ']' + cd .. + rm -rf es-index-cleaner-managed + warning 'es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + start_test es-multiinstance + '[' 1 -ne 1 ']' + test_name=es-multiinstance + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-multiinstance' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-multiinstance\e[0m' Rendering files for test es-multiinstance + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-multiinstance + cd es-multiinstance + jaeger_name=instance-1 + render_install_jaeger instance-1 production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=instance-1 + JAEGER_NAME=instance-1 + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f ./03-create-second-instance.yaml.template -o 03-create-second-instance.yaml + '[' true = true ']' + skip_test es-rollover-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-rollover-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-multiinstance + '[' es-multiinstance '!=' _build ']' + cd .. + rm -rf es-rollover-upstream + warning 'es-rollover-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_rollover -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-rollover-autoprov + '[' 1 -ne 1 ']' + test_name=es-rollover-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-rollover-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-rollover-autoprov\e[0m' Rendering files for test es-rollover-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-rollover-autoprov + cd es-rollover-autoprov + cp ../../es-rollover-upstream/05-assert.yaml ../../es-rollover-upstream/05-install.yaml ../../es-rollover-upstream/README.md . + jaeger_name=my-jaeger + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_report_spans my-jaeger true 2 00 true 02 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 00 03 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=00 + test_step=03 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./03-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./03-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 01 04 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=01 + test_step=04 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=01 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./04-check-indices.yaml + JOB_NUMBER=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./04-assert.yaml + render_report_spans my-jaeger true 2 02 true 06 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=02 + ensure_reported_spans=true + test_step=06 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=02 + JOB_NUMBER=02 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./06-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./06-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 02 07 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=02 + test_step=07 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=02 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./07-check-indices.yaml + JOB_NUMBER=02 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./07-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' 03 08 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + job_number=03 + test_step=08 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=03 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./08-check-indices.yaml + JOB_NUMBER=03 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./08-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' 04 09 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + job_number=04 + test_step=09 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=04 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./09-check-indices.yaml + JOB_NUMBER=04 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./09-assert.yaml + render_report_spans my-jaeger true 2 03 true 10 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=03 + ensure_reported_spans=true + test_step=10 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=03 + JOB_NUMBER=03 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./10-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./10-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + CRONJOB_NAME=my-jaeger-es-rollover + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./11-wait-rollover.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-000002'\'',' 05 11 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-000002'\'',' + job_number=05 + test_step=11 + escape_command ''\''--name'\'', '\''jaeger-span-000002'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-000002'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-000002'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-000002'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=05 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./11-check-indices.yaml + JOB_NUMBER=05 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./11-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' 06 12 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + job_number=06 + test_step=12 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=06 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./12-check-indices.yaml + JOB_NUMBER=06 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./12-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.3"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.3"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.15},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.3"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.3 ++ version_ge 5.8.3 5.4 +++ echo 5.8.3 5.4 +++ tr ' ' '\n' +++ sort -rV +++ head -n 1 ++ test 5.8.3 == 5.8.3 + '[' -n '' ']' + skip_test es-rollover-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-rollover-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-rollover-autoprov + '[' es-rollover-autoprov '!=' _build ']' + cd .. + rm -rf es-rollover-managed + warning 'es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + skip_test es-spark-dependencies 'This test is not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=es-spark-dependencies + message='This test is not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + rm -rf es-spark-dependencies + warning 'es-spark-dependencies: This test is not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-spark-dependencies: This test is not supported in OpenShift\e[0m' WAR: es-spark-dependencies: This test is not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running elasticsearch E2E tests' Running elasticsearch E2E tests + cd tests/e2e/elasticsearch/_build + set +e + KUBECONFIG=/tmp/kubeconfig-559878434 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 7 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/es-from-aio-to-production === PAUSE kuttl/harness/es-from-aio-to-production === RUN kuttl/harness/es-increasing-replicas === PAUSE kuttl/harness/es-increasing-replicas === RUN kuttl/harness/es-index-cleaner-autoprov === PAUSE kuttl/harness/es-index-cleaner-autoprov === RUN kuttl/harness/es-multiinstance === PAUSE kuttl/harness/es-multiinstance === RUN kuttl/harness/es-rollover-autoprov === PAUSE kuttl/harness/es-rollover-autoprov === RUN kuttl/harness/es-simple-prod === PAUSE kuttl/harness/es-simple-prod === CONT kuttl/harness/artifacts logger.go:42: 07:56:49 | artifacts | Creating namespace: kuttl-test-powerful-sunbird logger.go:42: 07:56:49 | artifacts | artifacts events from ns kuttl-test-powerful-sunbird: logger.go:42: 07:56:49 | artifacts | Deleting namespace: kuttl-test-powerful-sunbird === CONT kuttl/harness/es-multiinstance logger.go:42: 07:56:55 | es-multiinstance | Ignoring 03-create-second-instance.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:56:55 | es-multiinstance | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:56:55 | es-multiinstance | Creating namespace: kuttl-test-singular-perch logger.go:42: 07:56:55 | es-multiinstance/0-clear-namespace | starting test step 0-clear-namespace logger.go:42: 07:56:55 | es-multiinstance/0-clear-namespace | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --ignore-not-found=true] logger.go:42: 07:56:55 | es-multiinstance/0-clear-namespace | test step completed 0-clear-namespace logger.go:42: 07:56:55 | es-multiinstance/1-install | starting test step 1-install logger.go:42: 07:56:55 | es-multiinstance/1-install | Jaeger:kuttl-test-singular-perch/instance-1 created logger.go:42: 07:57:31 | es-multiinstance/1-install | test step completed 1-install logger.go:42: 07:57:31 | es-multiinstance/2-create-namespace | starting test step 2-create-namespace logger.go:42: 07:57:31 | es-multiinstance/2-create-namespace | running command: [sh -c kubectl create namespace jaeger-e2e-multiinstance-test] logger.go:42: 07:57:31 | es-multiinstance/2-create-namespace | namespace/jaeger-e2e-multiinstance-test created logger.go:42: 07:57:31 | es-multiinstance/2-create-namespace | test step completed 2-create-namespace logger.go:42: 07:57:31 | es-multiinstance/3-create-second-instance | starting test step 3-create-second-instance logger.go:42: 07:57:31 | es-multiinstance/3-create-second-instance | running command: [sh -c kubectl apply -f ./01-install.yaml -n jaeger-e2e-multiinstance-test] logger.go:42: 07:57:32 | es-multiinstance/3-create-second-instance | jaeger.jaegertracing.io/instance-1 created logger.go:42: 07:57:32 | es-multiinstance/3-create-second-instance | running command: [sh -c /tmp/jaeger-tests/bin/kubectl-kuttl assert ./01-assert.yaml -n jaeger-e2e-multiinstance-test --timeout 1000] logger.go:42: 07:58:21 | es-multiinstance/3-create-second-instance | assert is valid logger.go:42: 07:58:21 | es-multiinstance/3-create-second-instance | test step completed 3-create-second-instance logger.go:42: 07:58:21 | es-multiinstance/4-check-secrets | starting test step 4-check-secrets logger.go:42: 07:58:21 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n $NAMESPACE > secret1] logger.go:42: 07:58:21 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n jaeger-e2e-multiinstance-test > secret2] logger.go:42: 07:58:21 | es-multiinstance/4-check-secrets | running command: [sh -c cmp --silent secret1 secret2 || exit 0] logger.go:42: 07:58:21 | es-multiinstance/4-check-secrets | test step completed 4-check-secrets logger.go:42: 07:58:21 | es-multiinstance/5-delete | starting test step 5-delete logger.go:42: 07:58:21 | es-multiinstance/5-delete | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --wait=false] logger.go:42: 07:58:21 | es-multiinstance/5-delete | namespace "jaeger-e2e-multiinstance-test" deleted logger.go:42: 07:58:21 | es-multiinstance/5-delete | test step completed 5-delete logger.go:42: 07:58:21 | es-multiinstance | es-multiinstance events from ns kuttl-test-singular-perch: logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:01 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestsingularperchinstance1-1-7f899b8cc6 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestsingularperchinstance1-1-7f899bc78v4 replicaset-controller logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:01 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsingularperchinstance1-1-7f899bc78v4 Binding Scheduled Successfully assigned kuttl-test-singular-perch/elasticsearch-cdm-kuttltestsingularperchinstance1-1-7f899bc78v4 to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:01 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestsingularperchinstance1-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestsingularperchinstance1-1-7f899b8cc6 to 1 deployment-controller logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:02 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsingularperchinstance1-1-7f899bc78v4 AddedInterface Add eth0 [10.129.2.21/23] from ovn-kubernetes multus logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:02 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsingularperchinstance1-1-7f899bc78v4.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:02 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsingularperchinstance1-1-7f899bc78v4.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:02 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsingularperchinstance1-1-7f899bc78v4.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:02 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsingularperchinstance1-1-7f899bc78v4.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:02 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsingularperchinstance1-1-7f899bc78v4.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:02 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsingularperchinstance1-1-7f899bc78v4.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:12 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestsingularperchinstance1-1-7f899bc78v4.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:17 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestsingularperchinstance1-1-7f899bc78v4.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:28 +0000 UTC Normal Pod instance-1-collector-57d58f498b-4nsdf Binding Scheduled Successfully assigned kuttl-test-singular-perch/instance-1-collector-57d58f498b-4nsdf to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:28 +0000 UTC Normal ReplicaSet.apps instance-1-collector-57d58f498b SuccessfulCreate Created pod: instance-1-collector-57d58f498b-4nsdf replicaset-controller logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:28 +0000 UTC Normal Deployment.apps instance-1-collector ScalingReplicaSet Scaled up replica set instance-1-collector-57d58f498b to 1 deployment-controller logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:28 +0000 UTC Normal Pod instance-1-query-68dd4f487-b8w5r Binding Scheduled Successfully assigned kuttl-test-singular-perch/instance-1-query-68dd4f487-b8w5r to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:28 +0000 UTC Normal ReplicaSet.apps instance-1-query-68dd4f487 SuccessfulCreate Created pod: instance-1-query-68dd4f487-b8w5r replicaset-controller logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:28 +0000 UTC Normal Deployment.apps instance-1-query ScalingReplicaSet Scaled up replica set instance-1-query-68dd4f487 to 1 deployment-controller logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:29 +0000 UTC Normal Pod instance-1-collector-57d58f498b-4nsdf AddedInterface Add eth0 [10.131.0.30/23] from ovn-kubernetes multus logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:29 +0000 UTC Normal Pod instance-1-collector-57d58f498b-4nsdf.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:261a2c64c771323ad259e14802703bb790ee6b486b72c4939b487fbd83c30703" already present on machine kubelet logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:29 +0000 UTC Normal Pod instance-1-collector-57d58f498b-4nsdf.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:29 +0000 UTC Normal Pod instance-1-collector-57d58f498b-4nsdf.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:29 +0000 UTC Normal Pod instance-1-query-68dd4f487-b8w5r AddedInterface Add eth0 [10.128.2.36/23] from ovn-kubernetes multus logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:29 +0000 UTC Normal Pod instance-1-query-68dd4f487-b8w5r.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:156ea63cbd4dcd8c44173f118dba74388adbf1e82493237a4c514bb14131d5b8" already present on machine kubelet logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:29 +0000 UTC Normal Pod instance-1-query-68dd4f487-b8w5r.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:29 +0000 UTC Normal Pod instance-1-query-68dd4f487-b8w5r.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:29 +0000 UTC Normal Pod instance-1-query-68dd4f487-b8w5r.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:29 +0000 UTC Normal Pod instance-1-query-68dd4f487-b8w5r.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:29 +0000 UTC Normal Pod instance-1-query-68dd4f487-b8w5r.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:29 +0000 UTC Normal Pod instance-1-query-68dd4f487-b8w5r.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" already present on machine kubelet logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:29 +0000 UTC Normal Pod instance-1-query-68dd4f487-b8w5r.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:29 +0000 UTC Normal Pod instance-1-query-68dd4f487-b8w5r.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:43 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:43 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:57:43 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:58:13 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod instance-1-collector-57d58f498b-4nsdf horizontal-pod-autoscaler logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:58:13 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod instance-1-collector-57d58f498b-4nsdf horizontal-pod-autoscaler logger.go:42: 07:58:21 | es-multiinstance | 2024-05-20 07:58:14 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod instance-1-collector-57d58f498b-4nsdf horizontal-pod-autoscaler logger.go:42: 07:58:21 | es-multiinstance | Deleting namespace: kuttl-test-singular-perch === CONT kuttl/harness/es-increasing-replicas logger.go:42: 07:58:28 | es-increasing-replicas | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:58:28 | es-increasing-replicas | Ignoring check-es-nodes.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:58:28 | es-increasing-replicas | Ignoring openshift-check-es-nodes.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:58:28 | es-increasing-replicas | Creating namespace: kuttl-test-desired-martin logger.go:42: 07:58:29 | es-increasing-replicas/1-install | starting test step 1-install logger.go:42: 07:58:29 | es-increasing-replicas/1-install | Jaeger:kuttl-test-desired-martin/simple-prod created logger.go:42: 07:59:06 | es-increasing-replicas/1-install | test step completed 1-install logger.go:42: 07:59:06 | es-increasing-replicas/2-install | starting test step 2-install logger.go:42: 07:59:06 | es-increasing-replicas/2-install | Jaeger:kuttl-test-desired-martin/simple-prod updated logger.go:42: 07:59:21 | es-increasing-replicas/2-install | test step completed 2-install logger.go:42: 07:59:21 | es-increasing-replicas/3-smoke-test | starting test step 3-smoke-test logger.go:42: 07:59:21 | es-increasing-replicas/3-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:59:23 | es-increasing-replicas/3-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:59:29 | es-increasing-replicas/3-smoke-test | running command: [sh -c ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337 JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:59:30 | es-increasing-replicas/3-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:59:30 | es-increasing-replicas/3-smoke-test | job.batch/report-span created logger.go:42: 07:59:31 | es-increasing-replicas/3-smoke-test | job.batch/check-span created logger.go:42: 07:59:42 | es-increasing-replicas/3-smoke-test | test step completed 3-smoke-test logger.go:42: 07:59:42 | es-increasing-replicas/4-install | starting test step 4-install logger.go:42: 07:59:42 | es-increasing-replicas/4-install | Jaeger:kuttl-test-desired-martin/simple-prod updated logger.go:42: 07:59:42 | es-increasing-replicas/4-install | test step completed 4-install logger.go:42: 07:59:42 | es-increasing-replicas/5-check-es-nodes | starting test step 5-check-es-nodes logger.go:42: 07:59:42 | es-increasing-replicas/5-check-es-nodes | running command: [sh -c ./check-es-nodes.sh $NAMESPACE] logger.go:42: 07:59:42 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 07:59:42 | es-increasing-replicas/5-check-es-nodes | false logger.go:42: 07:59:42 | es-increasing-replicas/5-check-es-nodes | Error: no matches found logger.go:42: 07:59:47 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 07:59:47 | es-increasing-replicas/5-check-es-nodes | true logger.go:42: 07:59:47 | es-increasing-replicas/5-check-es-nodes | test step completed 5-check-es-nodes logger.go:42: 07:59:47 | es-increasing-replicas | es-increasing-replicas events from ns kuttl-test-desired-martin: logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:58:35 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestdesiredmartinsimpleprod-1-7c894c879b SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestdesiredmartinsimpleprod-1-7c894m7xzg replicaset-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:58:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdesiredmartinsimpleprod-1-7c894m7xzg Binding Scheduled Successfully assigned kuttl-test-desired-martin/elasticsearch-cdm-kuttltestdesiredmartinsimpleprod-1-7c894m7xzg to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:58:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdesiredmartinsimpleprod-1-7c894m7xzg AddedInterface Add eth0 [10.129.2.22/23] from ovn-kubernetes multus logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:58:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdesiredmartinsimpleprod-1-7c894m7xzg.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:58:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdesiredmartinsimpleprod-1-7c894m7xzg.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:58:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdesiredmartinsimpleprod-1-7c894m7xzg.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:58:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdesiredmartinsimpleprod-1-7c894m7xzg.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:58:35 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestdesiredmartinsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestdesiredmartinsimpleprod-1-7c894c879b to 1 deployment-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:58:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdesiredmartinsimpleprod-1-7c894m7xzg.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:58:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdesiredmartinsimpleprod-1-7c894m7xzg.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:58:45 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestdesiredmartinsimpleprod-1-7c894m7xzg.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:58:50 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestdesiredmartinsimpleprod-1-7c894m7xzg.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:02 +0000 UTC Normal Pod simple-prod-collector-5bffdbfd5d-7vfwj Binding Scheduled Successfully assigned kuttl-test-desired-martin/simple-prod-collector-5bffdbfd5d-7vfwj to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:02 +0000 UTC Normal Pod simple-prod-collector-5bffdbfd5d-7vfwj AddedInterface Add eth0 [10.128.2.39/23] from ovn-kubernetes multus logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:02 +0000 UTC Normal Pod simple-prod-collector-5bffdbfd5d-7vfwj.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:261a2c64c771323ad259e14802703bb790ee6b486b72c4939b487fbd83c30703" already present on machine kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:02 +0000 UTC Normal Pod simple-prod-collector-5bffdbfd5d-7vfwj.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:02 +0000 UTC Normal Pod simple-prod-collector-5bffdbfd5d-7vfwj.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:02 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-5bffdbfd5d SuccessfulCreate Created pod: simple-prod-collector-5bffdbfd5d-7vfwj replicaset-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:02 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-5bffdbfd5d to 1 deployment-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:02 +0000 UTC Normal Pod simple-prod-query-6b88df45b5-w62z2 Binding Scheduled Successfully assigned kuttl-test-desired-martin/simple-prod-query-6b88df45b5-w62z2 to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:02 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6b88df45b5 SuccessfulCreate Created pod: simple-prod-query-6b88df45b5-w62z2 replicaset-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:02 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-6b88df45b5 to 1 deployment-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:03 +0000 UTC Normal Pod simple-prod-query-6b88df45b5-w62z2 AddedInterface Add eth0 [10.128.2.40/23] from ovn-kubernetes multus logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:03 +0000 UTC Normal Pod simple-prod-query-6b88df45b5-w62z2.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:156ea63cbd4dcd8c44173f118dba74388adbf1e82493237a4c514bb14131d5b8" already present on machine kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:03 +0000 UTC Normal Pod simple-prod-query-6b88df45b5-w62z2.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:03 +0000 UTC Normal Pod simple-prod-query-6b88df45b5-w62z2.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:03 +0000 UTC Normal Pod simple-prod-query-6b88df45b5-w62z2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:03 +0000 UTC Normal Pod simple-prod-query-6b88df45b5-w62z2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:03 +0000 UTC Normal Pod simple-prod-query-6b88df45b5-w62z2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:03 +0000 UTC Normal Pod simple-prod-query-6b88df45b5-w62z2.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" already present on machine kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:04 +0000 UTC Normal Pod simple-prod-query-6b88df45b5-w62z2.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:04 +0000 UTC Normal Pod simple-prod-query-6b88df45b5-w62z2.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:09 +0000 UTC Normal Pod simple-prod-collector-5bffdbfd5d-r9pfl Binding Scheduled Successfully assigned kuttl-test-desired-martin/simple-prod-collector-5bffdbfd5d-r9pfl to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:09 +0000 UTC Normal Pod simple-prod-collector-5bffdbfd5d-r9pfl AddedInterface Add eth0 [10.131.0.32/23] from ovn-kubernetes multus logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:09 +0000 UTC Normal Pod simple-prod-collector-5bffdbfd5d-r9pfl.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:261a2c64c771323ad259e14802703bb790ee6b486b72c4939b487fbd83c30703" already present on machine kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:09 +0000 UTC Normal Pod simple-prod-collector-5bffdbfd5d-r9pfl.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:09 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-5bffdbfd5d SuccessfulCreate Created pod: simple-prod-collector-5bffdbfd5d-r9pfl replicaset-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:09 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-5bffdbfd5d to 2 from 1 deployment-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:09 +0000 UTC Normal Pod simple-prod-query-6b88df45b5-db98f Binding Scheduled Successfully assigned kuttl-test-desired-martin/simple-prod-query-6b88df45b5-db98f to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:09 +0000 UTC Normal Pod simple-prod-query-6b88df45b5-db98f AddedInterface Add eth0 [10.131.0.33/23] from ovn-kubernetes multus logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:09 +0000 UTC Normal Pod simple-prod-query-6b88df45b5-db98f.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:156ea63cbd4dcd8c44173f118dba74388adbf1e82493237a4c514bb14131d5b8" kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:09 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6b88df45b5 SuccessfulCreate Created pod: simple-prod-query-6b88df45b5-db98f replicaset-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:09 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-6b88df45b5 to 2 from 1 deployment-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:10 +0000 UTC Normal Pod simple-prod-collector-5bffdbfd5d-r9pfl.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:12 +0000 UTC Normal Pod simple-prod-query-6b88df45b5-db98f.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:156ea63cbd4dcd8c44173f118dba74388adbf1e82493237a4c514bb14131d5b8" in 2.125s (2.125s including waiting) kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:12 +0000 UTC Normal Pod simple-prod-query-6b88df45b5-db98f.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:12 +0000 UTC Normal Pod simple-prod-query-6b88df45b5-db98f.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:12 +0000 UTC Normal Pod simple-prod-query-6b88df45b5-db98f.spec.containers{oauth-proxy} Pulling Pulling image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:17 +0000 UTC Normal Pod simple-prod-query-6b88df45b5-db98f.spec.containers{oauth-proxy} Pulled Successfully pulled image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" in 4.871s (4.871s including waiting) kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:17 +0000 UTC Normal Pod simple-prod-query-6b88df45b5-db98f.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:17 +0000 UTC Normal Pod simple-prod-query-6b88df45b5-db98f.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:17 +0000 UTC Normal Pod simple-prod-query-6b88df45b5-db98f.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:18 +0000 UTC Normal Pod simple-prod-query-6b88df45b5-db98f.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" in 1.751s (1.751s including waiting) kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:19 +0000 UTC Normal Pod simple-prod-query-6b88df45b5-db98f.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:19 +0000 UTC Normal Pod simple-prod-query-6b88df45b5-db98f.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:24 +0000 UTC Normal Pod simple-prod-query-6b88df45b5-db98f.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:24 +0000 UTC Normal Pod simple-prod-query-6b88df45b5-db98f.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:24 +0000 UTC Normal Pod simple-prod-query-6b88df45b5-db98f.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:24 +0000 UTC Normal Pod simple-prod-query-6b88df45b5-w62z2.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:24 +0000 UTC Normal Pod simple-prod-query-6b88df45b5-w62z2.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:24 +0000 UTC Normal Pod simple-prod-query-6b88df45b5-w62z2.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:24 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6b88df45b5 SuccessfulDelete Deleted pod: simple-prod-query-6b88df45b5-db98f replicaset-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:24 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6b88df45b5 SuccessfulDelete Deleted pod: simple-prod-query-6b88df45b5-w62z2 replicaset-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:24 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-6b88df45b5 to 0 from 2 deployment-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:25 +0000 UTC Normal Pod simple-prod-query-6856649d8f-64p4m Binding Scheduled Successfully assigned kuttl-test-desired-martin/simple-prod-query-6856649d8f-64p4m to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:25 +0000 UTC Normal Pod simple-prod-query-6856649d8f-9t9ms Binding Scheduled Successfully assigned kuttl-test-desired-martin/simple-prod-query-6856649d8f-9t9ms to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:25 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6856649d8f SuccessfulCreate Created pod: simple-prod-query-6856649d8f-9t9ms replicaset-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:25 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6856649d8f SuccessfulCreate Created pod: simple-prod-query-6856649d8f-64p4m replicaset-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:25 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-6856649d8f to 2 deployment-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:26 +0000 UTC Normal Pod simple-prod-query-6856649d8f-64p4m AddedInterface Add eth0 [10.131.0.34/23] from ovn-kubernetes multus logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:26 +0000 UTC Normal Pod simple-prod-query-6856649d8f-64p4m.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:156ea63cbd4dcd8c44173f118dba74388adbf1e82493237a4c514bb14131d5b8" already present on machine kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:26 +0000 UTC Normal Pod simple-prod-query-6856649d8f-64p4m.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:26 +0000 UTC Normal Pod simple-prod-query-6856649d8f-64p4m.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:26 +0000 UTC Normal Pod simple-prod-query-6856649d8f-64p4m.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:26 +0000 UTC Normal Pod simple-prod-query-6856649d8f-64p4m.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:26 +0000 UTC Normal Pod simple-prod-query-6856649d8f-64p4m.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:26 +0000 UTC Normal Pod simple-prod-query-6856649d8f-64p4m.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" already present on machine kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:26 +0000 UTC Normal Pod simple-prod-query-6856649d8f-64p4m.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:26 +0000 UTC Normal Pod simple-prod-query-6856649d8f-64p4m.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:26 +0000 UTC Normal Pod simple-prod-query-6856649d8f-9t9ms AddedInterface Add eth0 [10.128.2.41/23] from ovn-kubernetes multus logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:26 +0000 UTC Normal Pod simple-prod-query-6856649d8f-9t9ms.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:156ea63cbd4dcd8c44173f118dba74388adbf1e82493237a4c514bb14131d5b8" already present on machine kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:26 +0000 UTC Normal Pod simple-prod-query-6856649d8f-9t9ms.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:26 +0000 UTC Normal Pod simple-prod-query-6856649d8f-9t9ms.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:26 +0000 UTC Normal Pod simple-prod-query-6856649d8f-9t9ms.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:26 +0000 UTC Normal Pod simple-prod-query-6856649d8f-9t9ms.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:26 +0000 UTC Normal Pod simple-prod-query-6856649d8f-9t9ms.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:26 +0000 UTC Normal Pod simple-prod-query-6856649d8f-9t9ms.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" already present on machine kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:26 +0000 UTC Normal Pod simple-prod-query-6856649d8f-9t9ms.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:26 +0000 UTC Normal Pod simple-prod-query-6856649d8f-9t9ms.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:30 +0000 UTC Normal Pod report-span-4fjvw Binding Scheduled Successfully assigned kuttl-test-desired-martin/report-span-4fjvw to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:30 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-4fjvw job-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:31 +0000 UTC Normal Pod check-span-nkmbw Binding Scheduled Successfully assigned kuttl-test-desired-martin/check-span-nkmbw to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:31 +0000 UTC Normal Pod check-span-nkmbw AddedInterface Add eth0 [10.131.0.36/23] from ovn-kubernetes multus logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:31 +0000 UTC Normal Pod check-span-nkmbw.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:31 +0000 UTC Normal Pod check-span-nkmbw.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:31 +0000 UTC Normal Pod check-span-nkmbw.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:31 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-nkmbw job-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:31 +0000 UTC Normal Pod report-span-4fjvw AddedInterface Add eth0 [10.131.0.35/23] from ovn-kubernetes multus logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:31 +0000 UTC Normal Pod report-span-4fjvw.spec.containers{report-span} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:31 +0000 UTC Normal Pod report-span-4fjvw.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:31 +0000 UTC Normal Pod report-span-4fjvw.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:41 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:43 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestdesiredmartinsimpleprod-2-6b78f454b4 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestdesiredmartinsimpleprod-2-6b78fst5cv replicaset-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdesiredmartinsimpleprod-2-6b78fst5cv Binding Scheduled Successfully assigned kuttl-test-desired-martin/elasticsearch-cdm-kuttltestdesiredmartinsimpleprod-2-6b78fst5cv to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:43 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestdesiredmartinsimpleprod-2 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestdesiredmartinsimpleprod-2-6b78f454b4 to 1 deployment-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:43 +0000 UTC Normal Pod simple-prod-collector-5bffdbfd5d-7vfwj.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:43 +0000 UTC Normal Pod simple-prod-collector-5bffdbfd5d-r9pfl.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:43 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-5bffdbfd5d SuccessfulDelete Deleted pod: simple-prod-collector-5bffdbfd5d-7vfwj replicaset-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:43 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-5bffdbfd5d SuccessfulDelete Deleted pod: simple-prod-collector-5bffdbfd5d-r9pfl replicaset-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:43 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled down replica set simple-prod-collector-5bffdbfd5d to 0 from 2 deployment-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:43 +0000 UTC Normal Pod simple-prod-query-6856649d8f-64p4m.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:43 +0000 UTC Normal Pod simple-prod-query-6856649d8f-64p4m.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:43 +0000 UTC Normal Pod simple-prod-query-6856649d8f-64p4m.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:43 +0000 UTC Normal Pod simple-prod-query-6856649d8f-9t9ms.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:43 +0000 UTC Normal Pod simple-prod-query-6856649d8f-9t9ms.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:43 +0000 UTC Normal Pod simple-prod-query-6856649d8f-9t9ms.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:43 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6856649d8f SuccessfulDelete Deleted pod: simple-prod-query-6856649d8f-64p4m replicaset-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:43 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6856649d8f SuccessfulDelete Deleted pod: simple-prod-query-6856649d8f-9t9ms replicaset-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:43 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-6856649d8f to 0 from 2 deployment-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:44 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdesiredmartinsimpleprod-2-6b78fst5cv AddedInterface Add eth0 [10.131.0.37/23] from ovn-kubernetes multus logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:44 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdesiredmartinsimpleprod-2-6b78fst5cv.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:44 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdesiredmartinsimpleprod-2-6b78fst5cv.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:44 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdesiredmartinsimpleprod-2-6b78fst5cv.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:44 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdesiredmartinsimpleprod-2-6b78fst5cv.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:44 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdesiredmartinsimpleprod-2-6b78fst5cv.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:44 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdesiredmartinsimpleprod-2-6b78fst5cv.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:45 +0000 UTC Normal Pod simple-prod-collector-69685cdc69-2b8g4 Binding Scheduled Successfully assigned kuttl-test-desired-martin/simple-prod-collector-69685cdc69-2b8g4 to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:45 +0000 UTC Normal Pod simple-prod-collector-69685cdc69-2b8g4 AddedInterface Add eth0 [10.131.0.38/23] from ovn-kubernetes multus logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:45 +0000 UTC Normal Pod simple-prod-collector-69685cdc69-2b8g4.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:261a2c64c771323ad259e14802703bb790ee6b486b72c4939b487fbd83c30703" already present on machine kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:45 +0000 UTC Normal Pod simple-prod-collector-69685cdc69-2b8g4.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:45 +0000 UTC Normal Pod simple-prod-collector-69685cdc69-772ws Binding Scheduled Successfully assigned kuttl-test-desired-martin/simple-prod-collector-69685cdc69-772ws to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:45 +0000 UTC Normal Pod simple-prod-collector-69685cdc69-772ws AddedInterface Add eth0 [10.128.2.42/23] from ovn-kubernetes multus logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:45 +0000 UTC Normal Pod simple-prod-collector-69685cdc69-772ws.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:261a2c64c771323ad259e14802703bb790ee6b486b72c4939b487fbd83c30703" already present on machine kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:45 +0000 UTC Normal Pod simple-prod-collector-69685cdc69-772ws.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:45 +0000 UTC Normal Pod simple-prod-collector-69685cdc69-772ws.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:45 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-69685cdc69 SuccessfulCreate Created pod: simple-prod-collector-69685cdc69-772ws replicaset-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:45 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-69685cdc69 SuccessfulCreate Created pod: simple-prod-collector-69685cdc69-2b8g4 replicaset-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:45 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-69685cdc69 to 2 deployment-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:45 +0000 UTC Normal Pod simple-prod-query-6c659dd768-52p22 Binding Scheduled Successfully assigned kuttl-test-desired-martin/simple-prod-query-6c659dd768-52p22 to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:45 +0000 UTC Normal Pod simple-prod-query-6c659dd768-52p22 AddedInterface Add eth0 [10.128.2.43/23] from ovn-kubernetes multus logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:45 +0000 UTC Normal Pod simple-prod-query-6c659dd768-52p22.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:156ea63cbd4dcd8c44173f118dba74388adbf1e82493237a4c514bb14131d5b8" already present on machine kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:45 +0000 UTC Normal Pod simple-prod-query-6c659dd768-52p22.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:45 +0000 UTC Normal Pod simple-prod-query-6c659dd768-52p22.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:45 +0000 UTC Normal Pod simple-prod-query-6c659dd768-52p22.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:45 +0000 UTC Normal Pod simple-prod-query-6c659dd768-pcm2h Binding Scheduled Successfully assigned kuttl-test-desired-martin/simple-prod-query-6c659dd768-pcm2h to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:45 +0000 UTC Normal Pod simple-prod-query-6c659dd768-pcm2h AddedInterface Add eth0 [10.131.0.39/23] from ovn-kubernetes multus logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:45 +0000 UTC Normal Pod simple-prod-query-6c659dd768-pcm2h.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:156ea63cbd4dcd8c44173f118dba74388adbf1e82493237a4c514bb14131d5b8" already present on machine kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:45 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6c659dd768 SuccessfulCreate Created pod: simple-prod-query-6c659dd768-52p22 replicaset-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:45 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6c659dd768 SuccessfulCreate Created pod: simple-prod-query-6c659dd768-pcm2h replicaset-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:45 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-6c659dd768 to 2 deployment-controller logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:46 +0000 UTC Normal Pod simple-prod-collector-69685cdc69-2b8g4.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:46 +0000 UTC Normal Pod simple-prod-query-6c659dd768-52p22.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:46 +0000 UTC Normal Pod simple-prod-query-6c659dd768-52p22.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:46 +0000 UTC Normal Pod simple-prod-query-6c659dd768-52p22.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" already present on machine kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:46 +0000 UTC Normal Pod simple-prod-query-6c659dd768-52p22.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:46 +0000 UTC Normal Pod simple-prod-query-6c659dd768-52p22.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:46 +0000 UTC Normal Pod simple-prod-query-6c659dd768-pcm2h.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:46 +0000 UTC Normal Pod simple-prod-query-6c659dd768-pcm2h.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:46 +0000 UTC Normal Pod simple-prod-query-6c659dd768-pcm2h.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:46 +0000 UTC Normal Pod simple-prod-query-6c659dd768-pcm2h.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:46 +0000 UTC Normal Pod simple-prod-query-6c659dd768-pcm2h.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:46 +0000 UTC Normal Pod simple-prod-query-6c659dd768-pcm2h.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" already present on machine kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:46 +0000 UTC Normal Pod simple-prod-query-6c659dd768-pcm2h.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:59:47 | es-increasing-replicas | 2024-05-20 07:59:46 +0000 UTC Normal Pod simple-prod-query-6c659dd768-pcm2h.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:59:47 | es-increasing-replicas | Deleting namespace: kuttl-test-desired-martin === CONT kuttl/harness/es-rollover-autoprov logger.go:42: 08:00:22 | es-rollover-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:00:22 | es-rollover-autoprov | Creating namespace: kuttl-test-united-stork logger.go:42: 08:00:22 | es-rollover-autoprov/1-install | starting test step 1-install logger.go:42: 08:00:22 | es-rollover-autoprov/1-install | Jaeger:kuttl-test-united-stork/my-jaeger created logger.go:42: 08:00:59 | es-rollover-autoprov/1-install | test step completed 1-install logger.go:42: 08:00:59 | es-rollover-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 08:00:59 | es-rollover-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 08:01:01 | es-rollover-autoprov/2-report-spans | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 08:01:09 | es-rollover-autoprov/2-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337 JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 08:01:10 | es-rollover-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 08:01:10 | es-rollover-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 08:01:35 | es-rollover-autoprov/2-report-spans | test step completed 2-report-spans logger.go:42: 08:01:35 | es-rollover-autoprov/3-check-indices | starting test step 3-check-indices logger.go:42: 08:01:35 | es-rollover-autoprov/3-check-indices | Job:kuttl-test-united-stork/00-check-indices created logger.go:42: 08:01:39 | es-rollover-autoprov/3-check-indices | test step completed 3-check-indices logger.go:42: 08:01:39 | es-rollover-autoprov/4-check-indices | starting test step 4-check-indices logger.go:42: 08:01:39 | es-rollover-autoprov/4-check-indices | Job:kuttl-test-united-stork/01-check-indices created logger.go:42: 08:01:43 | es-rollover-autoprov/4-check-indices | test step completed 4-check-indices logger.go:42: 08:01:43 | es-rollover-autoprov/5-install | starting test step 5-install logger.go:42: 08:01:43 | es-rollover-autoprov/5-install | Jaeger:kuttl-test-united-stork/my-jaeger updated logger.go:42: 08:01:56 | es-rollover-autoprov/5-install | test step completed 5-install logger.go:42: 08:01:56 | es-rollover-autoprov/6-report-spans | starting test step 6-report-spans logger.go:42: 08:01:56 | es-rollover-autoprov/6-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 08:02:04 | es-rollover-autoprov/6-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337 JOB_NUMBER=02 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-02-job.yaml] logger.go:42: 08:02:04 | es-rollover-autoprov/6-report-spans | running command: [sh -c kubectl apply -f report-span-02-job.yaml -n $NAMESPACE] logger.go:42: 08:02:05 | es-rollover-autoprov/6-report-spans | job.batch/02-report-span created logger.go:42: 08:02:28 | es-rollover-autoprov/6-report-spans | test step completed 6-report-spans logger.go:42: 08:02:28 | es-rollover-autoprov/7-check-indices | starting test step 7-check-indices logger.go:42: 08:02:28 | es-rollover-autoprov/7-check-indices | Job:kuttl-test-united-stork/02-check-indices created logger.go:42: 08:02:32 | es-rollover-autoprov/7-check-indices | test step completed 7-check-indices logger.go:42: 08:02:32 | es-rollover-autoprov/8-check-indices | starting test step 8-check-indices logger.go:42: 08:02:32 | es-rollover-autoprov/8-check-indices | Job:kuttl-test-united-stork/03-check-indices created logger.go:42: 08:02:37 | es-rollover-autoprov/8-check-indices | test step completed 8-check-indices logger.go:42: 08:02:37 | es-rollover-autoprov/9-check-indices | starting test step 9-check-indices logger.go:42: 08:02:37 | es-rollover-autoprov/9-check-indices | Job:kuttl-test-united-stork/04-check-indices created logger.go:42: 08:02:41 | es-rollover-autoprov/9-check-indices | test step completed 9-check-indices logger.go:42: 08:02:41 | es-rollover-autoprov/10-report-spans | starting test step 10-report-spans logger.go:42: 08:02:41 | es-rollover-autoprov/10-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 08:02:49 | es-rollover-autoprov/10-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337 JOB_NUMBER=03 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-03-job.yaml] logger.go:42: 08:02:49 | es-rollover-autoprov/10-report-spans | running command: [sh -c kubectl apply -f report-span-03-job.yaml -n $NAMESPACE] logger.go:42: 08:02:50 | es-rollover-autoprov/10-report-spans | job.batch/03-report-span created logger.go:42: 08:03:13 | es-rollover-autoprov/10-report-spans | test step completed 10-report-spans logger.go:42: 08:03:13 | es-rollover-autoprov/11-check-indices | starting test step 11-check-indices logger.go:42: 08:03:13 | es-rollover-autoprov/11-check-indices | running command: [sh -c go run ../../../../cmd-utils/wait-cronjob/main.go --cronjob my-jaeger-es-rollover --namespace $NAMESPACE] logger.go:42: 08:03:23 | es-rollover-autoprov/11-check-indices | time="2024-05-20T08:03:23Z" level=debug msg="Checking if the my-jaeger-es-rollover CronJob exists" logger.go:42: 08:03:23 | es-rollover-autoprov/11-check-indices | time="2024-05-20T08:03:23Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 08:03:23 | es-rollover-autoprov/11-check-indices | time="2024-05-20T08:03:23Z" level=info msg="Cronjob my-jaeger-es-rollover found successfully" logger.go:42: 08:03:23 | es-rollover-autoprov/11-check-indices | time="2024-05-20T08:03:23Z" level=debug msg="Waiting for the next scheduled job from my-jaeger-es-rollover cronjob" logger.go:42: 08:03:23 | es-rollover-autoprov/11-check-indices | time="2024-05-20T08:03:23Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 08:03:33 | es-rollover-autoprov/11-check-indices | time="2024-05-20T08:03:33Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 08:03:43 | es-rollover-autoprov/11-check-indices | time="2024-05-20T08:03:43Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 08:03:53 | es-rollover-autoprov/11-check-indices | time="2024-05-20T08:03:53Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 08:04:03 | es-rollover-autoprov/11-check-indices | time="2024-05-20T08:04:03Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 08:04:13 | es-rollover-autoprov/11-check-indices | time="2024-05-20T08:04:13Z" level=info msg="Job of owner my-jaeger-es-rollover succeeded after my-jaeger-es-rollover 50.050605871s" logger.go:42: 08:04:13 | es-rollover-autoprov/11-check-indices | Job:kuttl-test-united-stork/05-check-indices created logger.go:42: 08:04:17 | es-rollover-autoprov/11-check-indices | test step completed 11-check-indices logger.go:42: 08:04:17 | es-rollover-autoprov/12-check-indices | starting test step 12-check-indices logger.go:42: 08:04:17 | es-rollover-autoprov/12-check-indices | Job:kuttl-test-united-stork/06-check-indices created logger.go:42: 08:04:20 | es-rollover-autoprov/12-check-indices | test step completed 12-check-indices logger.go:42: 08:04:21 | es-rollover-autoprov | es-rollover-autoprov events from ns kuttl-test-united-stork: logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:28 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestunitedstorkmyjaeger-1-69654bcfdb SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestunitedstorkmyjaeger-1-69654bcfd8g86l replicaset-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:28 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestunitedstorkmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestunitedstorkmyjaeger-1-69654bcfdb to 1 deployment-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestunitedstorkmyjaeger-1-69654bcfd8g86l Binding Scheduled Successfully assigned kuttl-test-united-stork/elasticsearch-cdm-kuttltestunitedstorkmyjaeger-1-69654bcfd8g86l to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:30 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestunitedstorkmyjaeger-1-69654bcfd8g86l AddedInterface Add eth0 [10.129.2.23/23] from ovn-kubernetes multus logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:30 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestunitedstorkmyjaeger-1-69654bcfd8g86l.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:30 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestunitedstorkmyjaeger-1-69654bcfd8g86l.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:30 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestunitedstorkmyjaeger-1-69654bcfd8g86l.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:30 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestunitedstorkmyjaeger-1-69654bcfd8g86l.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:30 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestunitedstorkmyjaeger-1-69654bcfd8g86l.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:30 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestunitedstorkmyjaeger-1-69654bcfd8g86l.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:40 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestunitedstorkmyjaeger-1-69654bcfd8g86l.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:45 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestunitedstorkmyjaeger-1-69654bcfd8g86l.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:56 +0000 UTC Normal Pod my-jaeger-collector-59bdb75cf7-whcfr Binding Scheduled Successfully assigned kuttl-test-united-stork/my-jaeger-collector-59bdb75cf7-whcfr to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:56 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-59bdb75cf7 SuccessfulCreate Created pod: my-jaeger-collector-59bdb75cf7-whcfr replicaset-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:56 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-59bdb75cf7 to 1 deployment-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:56 +0000 UTC Normal Pod my-jaeger-query-5f7f77797d-28hqq Binding Scheduled Successfully assigned kuttl-test-united-stork/my-jaeger-query-5f7f77797d-28hqq to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:56 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-5f7f77797d SuccessfulCreate Created pod: my-jaeger-query-5f7f77797d-28hqq replicaset-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:56 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-5f7f77797d to 1 deployment-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:57 +0000 UTC Normal Pod my-jaeger-collector-59bdb75cf7-whcfr AddedInterface Add eth0 [10.131.0.40/23] from ovn-kubernetes multus logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:57 +0000 UTC Normal Pod my-jaeger-collector-59bdb75cf7-whcfr.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:261a2c64c771323ad259e14802703bb790ee6b486b72c4939b487fbd83c30703" already present on machine kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:57 +0000 UTC Normal Pod my-jaeger-collector-59bdb75cf7-whcfr.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:57 +0000 UTC Normal Pod my-jaeger-collector-59bdb75cf7-whcfr.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:57 +0000 UTC Normal Pod my-jaeger-query-5f7f77797d-28hqq AddedInterface Add eth0 [10.128.2.45/23] from ovn-kubernetes multus logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:57 +0000 UTC Normal Pod my-jaeger-query-5f7f77797d-28hqq.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:156ea63cbd4dcd8c44173f118dba74388adbf1e82493237a4c514bb14131d5b8" already present on machine kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:57 +0000 UTC Normal Pod my-jaeger-query-5f7f77797d-28hqq.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:57 +0000 UTC Normal Pod my-jaeger-query-5f7f77797d-28hqq.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:57 +0000 UTC Normal Pod my-jaeger-query-5f7f77797d-28hqq.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:57 +0000 UTC Normal Pod my-jaeger-query-5f7f77797d-28hqq.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:57 +0000 UTC Normal Pod my-jaeger-query-5f7f77797d-28hqq.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:57 +0000 UTC Normal Pod my-jaeger-query-5f7f77797d-28hqq.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" already present on machine kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:57 +0000 UTC Normal Pod my-jaeger-query-5f7f77797d-28hqq.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:00:57 +0000 UTC Normal Pod my-jaeger-query-5f7f77797d-28hqq.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:06 +0000 UTC Normal Pod my-jaeger-query-56f98bbf-mxjpl Binding Scheduled Successfully assigned kuttl-test-united-stork/my-jaeger-query-56f98bbf-mxjpl to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:06 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-56f98bbf SuccessfulCreate Created pod: my-jaeger-query-56f98bbf-mxjpl replicaset-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:06 +0000 UTC Normal Pod my-jaeger-query-5f7f77797d-28hqq.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:06 +0000 UTC Normal Pod my-jaeger-query-5f7f77797d-28hqq.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:06 +0000 UTC Normal Pod my-jaeger-query-5f7f77797d-28hqq.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:06 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-5f7f77797d SuccessfulDelete Deleted pod: my-jaeger-query-5f7f77797d-28hqq replicaset-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:06 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-5f7f77797d to 0 from 1 deployment-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:06 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-56f98bbf to 1 deployment-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:07 +0000 UTC Normal Pod my-jaeger-query-56f98bbf-mxjpl AddedInterface Add eth0 [10.128.2.46/23] from ovn-kubernetes multus logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:07 +0000 UTC Normal Pod my-jaeger-query-56f98bbf-mxjpl.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:156ea63cbd4dcd8c44173f118dba74388adbf1e82493237a4c514bb14131d5b8" already present on machine kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:07 +0000 UTC Normal Pod my-jaeger-query-56f98bbf-mxjpl.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:07 +0000 UTC Normal Pod my-jaeger-query-56f98bbf-mxjpl.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:07 +0000 UTC Normal Pod my-jaeger-query-56f98bbf-mxjpl.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:07 +0000 UTC Normal Pod my-jaeger-query-56f98bbf-mxjpl.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:07 +0000 UTC Normal Pod my-jaeger-query-56f98bbf-mxjpl.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:07 +0000 UTC Normal Pod my-jaeger-query-56f98bbf-mxjpl.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" already present on machine kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:07 +0000 UTC Normal Pod my-jaeger-query-56f98bbf-mxjpl.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:07 +0000 UTC Normal Pod my-jaeger-query-56f98bbf-mxjpl.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:10 +0000 UTC Normal Pod 00-report-span-lc8tr Binding Scheduled Successfully assigned kuttl-test-united-stork/00-report-span-lc8tr to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:10 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-lc8tr job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:11 +0000 UTC Normal Pod 00-report-span-lc8tr AddedInterface Add eth0 [10.131.0.41/23] from ovn-kubernetes multus logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:11 +0000 UTC Normal Pod 00-report-span-lc8tr.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:11 +0000 UTC Normal Pod 00-report-span-lc8tr.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:11 +0000 UTC Normal Pod 00-report-span-lc8tr.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:14 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:14 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:14 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:29 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:29 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-59bdb75cf7-whcfr horizontal-pod-autoscaler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:29 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:34 +0000 UTC Normal Job.batch 00-report-span Completed Job completed job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:35 +0000 UTC Normal Pod 00-check-indices-sqm9t Binding Scheduled Successfully assigned kuttl-test-united-stork/00-check-indices-sqm9t to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:35 +0000 UTC Normal Pod 00-check-indices-sqm9t AddedInterface Add eth0 [10.131.0.42/23] from ovn-kubernetes multus logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:35 +0000 UTC Normal Pod 00-check-indices-sqm9t.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:35 +0000 UTC Normal Pod 00-check-indices-sqm9t.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:35 +0000 UTC Normal Pod 00-check-indices-sqm9t.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:35 +0000 UTC Normal Job.batch 00-check-indices SuccessfulCreate Created pod: 00-check-indices-sqm9t job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:38 +0000 UTC Normal Job.batch 00-check-indices Completed Job completed job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:39 +0000 UTC Normal Pod 01-check-indices-rvwvp Binding Scheduled Successfully assigned kuttl-test-united-stork/01-check-indices-rvwvp to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:39 +0000 UTC Normal Pod 01-check-indices-rvwvp AddedInterface Add eth0 [10.131.0.43/23] from ovn-kubernetes multus logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:39 +0000 UTC Normal Job.batch 01-check-indices SuccessfulCreate Created pod: 01-check-indices-rvwvp job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:40 +0000 UTC Normal Pod 01-check-indices-rvwvp.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:40 +0000 UTC Normal Pod 01-check-indices-rvwvp.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:40 +0000 UTC Normal Pod 01-check-indices-rvwvp.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:42 +0000 UTC Normal Job.batch 01-check-indices Completed Job completed job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:44 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-59bdb75cf7-whcfr horizontal-pod-autoscaler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:44 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-59bdb75cf7-whcfr horizontal-pod-autoscaler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:44 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-xmmbx Binding Scheduled Successfully assigned kuttl-test-united-stork/my-jaeger-es-rollover-create-mapping-xmmbx to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:44 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping SuccessfulCreate Created pod: my-jaeger-es-rollover-create-mapping-xmmbx job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:45 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-xmmbx AddedInterface Add eth0 [10.131.0.44/23] from ovn-kubernetes multus logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:45 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-xmmbx.spec.containers{my-jaeger-es-rollover-create-mapping} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:251db17ea938680998301b56606181a4eae7cfb9325f5fb0d51497edb65cdc3a" kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:48 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-xmmbx.spec.containers{my-jaeger-es-rollover-create-mapping} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:251db17ea938680998301b56606181a4eae7cfb9325f5fb0d51497edb65cdc3a" in 2.828s (2.828s including waiting) kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:48 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-xmmbx.spec.containers{my-jaeger-es-rollover-create-mapping} Created Created container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:48 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-xmmbx.spec.containers{my-jaeger-es-rollover-create-mapping} Started Started container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:51 +0000 UTC Normal Pod my-jaeger-collector-59bdb75cf7-whcfr.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:51 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-59bdb75cf7 SuccessfulDelete Deleted pod: my-jaeger-collector-59bdb75cf7-whcfr replicaset-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:51 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-59bdb75cf7 to 0 from 1 deployment-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:51 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping Completed Job completed job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:51 +0000 UTC Normal Pod my-jaeger-query-56f98bbf-mxjpl.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:51 +0000 UTC Normal Pod my-jaeger-query-56f98bbf-mxjpl.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:51 +0000 UTC Normal Pod my-jaeger-query-56f98bbf-mxjpl.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:51 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-56f98bbf SuccessfulDelete Deleted pod: my-jaeger-query-56f98bbf-mxjpl replicaset-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:51 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-56f98bbf to 0 from 1 deployment-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:52 +0000 UTC Normal Pod my-jaeger-collector-7888b9bb4d-9dsfc Binding Scheduled Successfully assigned kuttl-test-united-stork/my-jaeger-collector-7888b9bb4d-9dsfc to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:52 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7888b9bb4d SuccessfulCreate Created pod: my-jaeger-collector-7888b9bb4d-9dsfc replicaset-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:52 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-7888b9bb4d to 1 deployment-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:52 +0000 UTC Normal Pod my-jaeger-query-8b5cbccf8-dbgz9 Binding Scheduled Successfully assigned kuttl-test-united-stork/my-jaeger-query-8b5cbccf8-dbgz9 to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:52 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-8b5cbccf8 SuccessfulCreate Created pod: my-jaeger-query-8b5cbccf8-dbgz9 replicaset-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:52 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-8b5cbccf8 to 1 deployment-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:53 +0000 UTC Normal Pod my-jaeger-collector-7888b9bb4d-9dsfc AddedInterface Add eth0 [10.131.0.45/23] from ovn-kubernetes multus logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:53 +0000 UTC Normal Pod my-jaeger-collector-7888b9bb4d-9dsfc.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:261a2c64c771323ad259e14802703bb790ee6b486b72c4939b487fbd83c30703" already present on machine kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:53 +0000 UTC Normal Pod my-jaeger-collector-7888b9bb4d-9dsfc.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:53 +0000 UTC Normal Pod my-jaeger-collector-7888b9bb4d-9dsfc.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:53 +0000 UTC Normal Pod my-jaeger-query-8b5cbccf8-dbgz9 AddedInterface Add eth0 [10.128.2.47/23] from ovn-kubernetes multus logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:53 +0000 UTC Normal Pod my-jaeger-query-8b5cbccf8-dbgz9.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:156ea63cbd4dcd8c44173f118dba74388adbf1e82493237a4c514bb14131d5b8" already present on machine kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:53 +0000 UTC Normal Pod my-jaeger-query-8b5cbccf8-dbgz9.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:53 +0000 UTC Normal Pod my-jaeger-query-8b5cbccf8-dbgz9.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:53 +0000 UTC Normal Pod my-jaeger-query-8b5cbccf8-dbgz9.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:53 +0000 UTC Normal Pod my-jaeger-query-8b5cbccf8-dbgz9.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:53 +0000 UTC Normal Pod my-jaeger-query-8b5cbccf8-dbgz9.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:53 +0000 UTC Normal Pod my-jaeger-query-8b5cbccf8-dbgz9.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" already present on machine kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:54 +0000 UTC Normal Pod my-jaeger-query-8b5cbccf8-dbgz9.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:01:54 +0000 UTC Normal Pod my-jaeger-query-8b5cbccf8-dbgz9.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28603202-v82s7 Binding Scheduled Successfully assigned kuttl-test-united-stork/my-jaeger-es-lookback-28603202-v82s7 to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28603202-v82s7 AddedInterface Add eth0 [10.131.0.47/23] from ovn-kubernetes multus logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28603202-v82s7.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:251db17ea938680998301b56606181a4eae7cfb9325f5fb0d51497edb65cdc3a" already present on machine kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28603202-v82s7.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28603202-v82s7.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28603202 SuccessfulCreate Created pod: my-jaeger-es-lookback-28603202-v82s7 job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28603202 cronjob-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28603202-46bg8 Binding Scheduled Successfully assigned kuttl-test-united-stork/my-jaeger-es-rollover-28603202-46bg8 to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28603202-46bg8 AddedInterface Add eth0 [10.131.0.46/23] from ovn-kubernetes multus logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28603202-46bg8.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:251db17ea938680998301b56606181a4eae7cfb9325f5fb0d51497edb65cdc3a" already present on machine kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28603202-46bg8.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28603202-46bg8.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28603202 SuccessfulCreate Created pod: my-jaeger-es-rollover-28603202-46bg8 job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28603202 cronjob-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:03 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28603202 Completed Job completed job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:03 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28603202, status: Complete cronjob-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:03 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28603202 Completed Job completed job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:03 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28603202, status: Complete cronjob-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:05 +0000 UTC Normal Pod 02-report-span-svrvb Binding Scheduled Successfully assigned kuttl-test-united-stork/02-report-span-svrvb to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:05 +0000 UTC Normal Pod 02-report-span-svrvb AddedInterface Add eth0 [10.131.0.48/23] from ovn-kubernetes multus logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:05 +0000 UTC Normal Pod 02-report-span-svrvb.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:05 +0000 UTC Normal Pod 02-report-span-svrvb.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:05 +0000 UTC Normal Pod 02-report-span-svrvb.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:05 +0000 UTC Normal Job.batch 02-report-span SuccessfulCreate Created pod: 02-report-span-svrvb job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:14 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-7888b9bb4d-9dsfc horizontal-pod-autoscaler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:28 +0000 UTC Normal Pod 02-check-indices-rc2c8 Binding Scheduled Successfully assigned kuttl-test-united-stork/02-check-indices-rc2c8 to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:28 +0000 UTC Normal Job.batch 02-check-indices SuccessfulCreate Created pod: 02-check-indices-rc2c8 job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:28 +0000 UTC Normal Job.batch 02-report-span Completed Job completed job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:29 +0000 UTC Normal Pod 02-check-indices-rc2c8 AddedInterface Add eth0 [10.131.0.49/23] from ovn-kubernetes multus logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:29 +0000 UTC Normal Pod 02-check-indices-rc2c8.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:29 +0000 UTC Normal Pod 02-check-indices-rc2c8.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:29 +0000 UTC Normal Pod 02-check-indices-rc2c8.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:29 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-7888b9bb4d-9dsfc horizontal-pod-autoscaler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:29 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-7888b9bb4d-9dsfc horizontal-pod-autoscaler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:32 +0000 UTC Normal Job.batch 02-check-indices Completed Job completed job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:32 +0000 UTC Normal Pod 03-check-indices-7sc96 Binding Scheduled Successfully assigned kuttl-test-united-stork/03-check-indices-7sc96 to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:32 +0000 UTC Normal Job.batch 03-check-indices SuccessfulCreate Created pod: 03-check-indices-7sc96 job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:33 +0000 UTC Normal Pod 03-check-indices-7sc96 AddedInterface Add eth0 [10.131.0.50/23] from ovn-kubernetes multus logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:33 +0000 UTC Normal Pod 03-check-indices-7sc96.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:33 +0000 UTC Normal Pod 03-check-indices-7sc96.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:33 +0000 UTC Normal Pod 03-check-indices-7sc96.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:36 +0000 UTC Normal Job.batch 03-check-indices Completed Job completed job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:37 +0000 UTC Normal Pod 04-check-indices-8k7j8 Binding Scheduled Successfully assigned kuttl-test-united-stork/04-check-indices-8k7j8 to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:37 +0000 UTC Normal Pod 04-check-indices-8k7j8 AddedInterface Add eth0 [10.131.0.51/23] from ovn-kubernetes multus logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:37 +0000 UTC Normal Pod 04-check-indices-8k7j8.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:37 +0000 UTC Normal Pod 04-check-indices-8k7j8.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:37 +0000 UTC Normal Pod 04-check-indices-8k7j8.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:37 +0000 UTC Normal Job.batch 04-check-indices SuccessfulCreate Created pod: 04-check-indices-8k7j8 job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:40 +0000 UTC Normal Job.batch 04-check-indices Completed Job completed job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:50 +0000 UTC Normal Pod 03-report-span-r9tm2 Binding Scheduled Successfully assigned kuttl-test-united-stork/03-report-span-r9tm2 to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:50 +0000 UTC Normal Pod 03-report-span-r9tm2 AddedInterface Add eth0 [10.131.0.52/23] from ovn-kubernetes multus logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:50 +0000 UTC Normal Pod 03-report-span-r9tm2.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:50 +0000 UTC Normal Pod 03-report-span-r9tm2.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:50 +0000 UTC Normal Pod 03-report-span-r9tm2.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:02:50 +0000 UTC Normal Job.batch 03-report-span SuccessfulCreate Created pod: 03-report-span-r9tm2 job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:03:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28603203-4rqn2 Binding Scheduled Successfully assigned kuttl-test-united-stork/my-jaeger-es-lookback-28603203-4rqn2 to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:03:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28603203-4rqn2 AddedInterface Add eth0 [10.131.0.54/23] from ovn-kubernetes multus logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:03:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28603203-4rqn2.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:251db17ea938680998301b56606181a4eae7cfb9325f5fb0d51497edb65cdc3a" already present on machine kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:03:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28603203-4rqn2.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:03:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28603203-4rqn2.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:03:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28603203 SuccessfulCreate Created pod: my-jaeger-es-lookback-28603203-4rqn2 job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:03:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28603203 cronjob-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:03:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28603203-84xg8 Binding Scheduled Successfully assigned kuttl-test-united-stork/my-jaeger-es-rollover-28603203-84xg8 to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:03:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28603203-84xg8 AddedInterface Add eth0 [10.131.0.53/23] from ovn-kubernetes multus logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:03:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28603203-84xg8.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:251db17ea938680998301b56606181a4eae7cfb9325f5fb0d51497edb65cdc3a" already present on machine kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:03:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28603203-84xg8.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:03:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28603203-84xg8.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:03:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28603203 SuccessfulCreate Created pod: my-jaeger-es-rollover-28603203-84xg8 job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:03:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28603203 cronjob-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:03:03 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28603203 Completed Job completed job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:03:03 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28603203, status: Complete cronjob-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:03:03 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28603203 Completed Job completed job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:03:03 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28603203, status: Complete cronjob-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:03:13 +0000 UTC Normal Job.batch 03-report-span Completed Job completed job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28603204-5m4x9 Binding Scheduled Successfully assigned kuttl-test-united-stork/my-jaeger-es-lookback-28603204-5m4x9 to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28603204-5m4x9 AddedInterface Add eth0 [10.131.0.55/23] from ovn-kubernetes multus logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28603204-5m4x9.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:251db17ea938680998301b56606181a4eae7cfb9325f5fb0d51497edb65cdc3a" already present on machine kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28603204-5m4x9.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28603204-5m4x9.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28603204 SuccessfulCreate Created pod: my-jaeger-es-lookback-28603204-5m4x9 job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28603204 cronjob-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28603204-pbss6 Binding Scheduled Successfully assigned kuttl-test-united-stork/my-jaeger-es-rollover-28603204-pbss6 to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28603204-pbss6 AddedInterface Add eth0 [10.131.0.56/23] from ovn-kubernetes multus logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28603204-pbss6.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:251db17ea938680998301b56606181a4eae7cfb9325f5fb0d51497edb65cdc3a" already present on machine kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28603204-pbss6.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28603204-pbss6.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28603204 SuccessfulCreate Created pod: my-jaeger-es-rollover-28603204-pbss6 job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28603204 cronjob-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:03 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28603204 Completed Job completed job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:03 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28603204, status: Complete cronjob-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:03 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28603204 Completed Job completed job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:03 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28603204, status: Complete cronjob-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:13 +0000 UTC Normal Pod 05-check-indices-q626b Binding Scheduled Successfully assigned kuttl-test-united-stork/05-check-indices-q626b to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:13 +0000 UTC Normal Job.batch 05-check-indices SuccessfulCreate Created pod: 05-check-indices-q626b job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:14 +0000 UTC Normal Pod 05-check-indices-q626b AddedInterface Add eth0 [10.131.0.57/23] from ovn-kubernetes multus logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:14 +0000 UTC Normal Pod 05-check-indices-q626b.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:14 +0000 UTC Normal Pod 05-check-indices-q626b.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:14 +0000 UTC Normal Pod 05-check-indices-q626b.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:16 +0000 UTC Normal Job.batch 05-check-indices Completed Job completed job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:17 +0000 UTC Normal Pod 06-check-indices-26f74 Binding Scheduled Successfully assigned kuttl-test-united-stork/06-check-indices-26f74 to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:17 +0000 UTC Normal Job.batch 06-check-indices SuccessfulCreate Created pod: 06-check-indices-26f74 job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:18 +0000 UTC Normal Pod 06-check-indices-26f74 AddedInterface Add eth0 [10.131.0.58/23] from ovn-kubernetes multus logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:18 +0000 UTC Normal Pod 06-check-indices-26f74.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:18 +0000 UTC Normal Pod 06-check-indices-26f74.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:18 +0000 UTC Normal Pod 06-check-indices-26f74.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:04:21 | es-rollover-autoprov | 2024-05-20 08:04:20 +0000 UTC Normal Job.batch 06-check-indices Completed Job completed job-controller logger.go:42: 08:04:21 | es-rollover-autoprov | Deleting namespace: kuttl-test-united-stork === CONT kuttl/harness/es-simple-prod logger.go:42: 08:04:28 | es-simple-prod | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:04:28 | es-simple-prod | Creating namespace: kuttl-test-bold-mullet logger.go:42: 08:04:28 | es-simple-prod | es-simple-prod events from ns kuttl-test-bold-mullet: logger.go:42: 08:04:28 | es-simple-prod | Deleting namespace: kuttl-test-bold-mullet === CONT kuttl/harness/es-index-cleaner-autoprov logger.go:42: 08:04:34 | es-index-cleaner-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:04:34 | es-index-cleaner-autoprov | Creating namespace: kuttl-test-maximum-jackal logger.go:42: 08:04:34 | es-index-cleaner-autoprov/1-install | starting test step 1-install logger.go:42: 08:04:34 | es-index-cleaner-autoprov/1-install | Jaeger:kuttl-test-maximum-jackal/test-es-index-cleaner-with-prefix created logger.go:42: 08:05:12 | es-index-cleaner-autoprov/1-install | test step completed 1-install logger.go:42: 08:05:12 | es-index-cleaner-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 08:05:12 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE test-es-index-cleaner-with-prefix /dev/null] logger.go:42: 08:05:14 | es-index-cleaner-autoprov/2-report-spans | Warning: resource jaegers/test-es-index-cleaner-with-prefix is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 08:05:20 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c DAYS=5 ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337 JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 08:05:21 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 08:05:21 | es-index-cleaner-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 08:06:00 | es-index-cleaner-autoprov/2-report-spans | test step completed 2-report-spans logger.go:42: 08:06:00 | es-index-cleaner-autoprov/3-install | starting test step 3-install logger.go:42: 08:06:00 | es-index-cleaner-autoprov/3-install | Jaeger:kuttl-test-maximum-jackal/test-es-index-cleaner-with-prefix updated logger.go:42: 08:06:00 | es-index-cleaner-autoprov/3-install | test step completed 3-install logger.go:42: 08:06:00 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | starting test step 4-wait-es-index-cleaner logger.go:42: 08:06:00 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | running command: [sh -c go run ../../../../cmd-utils/wait-cronjob/main.go --cronjob test-es-index-cleaner-with-prefix-es-index-cleaner --namespace $NAMESPACE] logger.go:42: 08:06:01 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-05-20T08:06:01Z" level=debug msg="Checking if the test-es-index-cleaner-with-prefix-es-index-cleaner CronJob exists" logger.go:42: 08:06:01 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-05-20T08:06:01Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 08:06:01 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-05-20T08:06:01Z" level=warning msg="The BatchV1/Cronjob test-es-index-cleaner-with-prefix-es-index-cleaner was not found" logger.go:42: 08:06:01 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-05-20T08:06:01Z" level=debug msg="Found BatchV/Cronjobs:" logger.go:42: 08:06:11 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-05-20T08:06:11Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 08:06:11 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-05-20T08:06:11Z" level=info msg="Cronjob test-es-index-cleaner-with-prefix-es-index-cleaner found successfully" logger.go:42: 08:06:11 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-05-20T08:06:11Z" level=debug msg="Waiting for the next scheduled job from test-es-index-cleaner-with-prefix-es-index-cleaner cronjob" logger.go:42: 08:06:11 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-05-20T08:06:11Z" level=debug msg="Waiting for next job from test-es-index-cleaner-with-prefix-es-index-cleaner to succeed" logger.go:42: 08:06:21 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-05-20T08:06:21Z" level=debug msg="Waiting for next job from test-es-index-cleaner-with-prefix-es-index-cleaner to succeed" logger.go:42: 08:06:31 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-05-20T08:06:31Z" level=debug msg="Waiting for next job from test-es-index-cleaner-with-prefix-es-index-cleaner to succeed" logger.go:42: 08:06:41 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-05-20T08:06:41Z" level=debug msg="Waiting for next job from test-es-index-cleaner-with-prefix-es-index-cleaner to succeed" logger.go:42: 08:06:51 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-05-20T08:06:51Z" level=debug msg="Waiting for next job from test-es-index-cleaner-with-prefix-es-index-cleaner to succeed" logger.go:42: 08:07:01 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-05-20T08:07:01Z" level=debug msg="Waiting for next job from test-es-index-cleaner-with-prefix-es-index-cleaner to succeed" logger.go:42: 08:07:11 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-05-20T08:07:11Z" level=info msg="Job of owner test-es-index-cleaner-with-prefix-es-index-cleaner succeeded after test-es-index-cleaner-with-prefix-es-index-cleaner 1m0.028040561s" logger.go:42: 08:07:11 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | test step completed 4-wait-es-index-cleaner logger.go:42: 08:07:11 | es-index-cleaner-autoprov/5-install | starting test step 5-install logger.go:42: 08:07:11 | es-index-cleaner-autoprov/5-install | Jaeger:kuttl-test-maximum-jackal/test-es-index-cleaner-with-prefix updated logger.go:42: 08:07:11 | es-index-cleaner-autoprov/5-install | test step completed 5-install logger.go:42: 08:07:11 | es-index-cleaner-autoprov/6-check-indices | starting test step 6-check-indices logger.go:42: 08:07:11 | es-index-cleaner-autoprov/6-check-indices | Job:kuttl-test-maximum-jackal/00-check-indices created logger.go:42: 08:07:15 | es-index-cleaner-autoprov/6-check-indices | test step completed 6-check-indices logger.go:42: 08:07:15 | es-index-cleaner-autoprov | es-index-cleaner-autoprov events from ns kuttl-test-maximum-jackal: logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:04:41 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmaximumjackaltestesindexcle-1-66v5rh Binding Scheduled Successfully assigned kuttl-test-maximum-jackal/elasticsearch-cdm-kuttltestmaximumjackaltestesindexcle-1-66v5rh to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:04:41 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestmaximumjackaltestesindexcle-1-68757df69d SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestmaximumjackaltestesindexcle-1-66v5rh replicaset-controller logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:04:41 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestmaximumjackaltestesindexcle-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestmaximumjackaltestesindexcle-1-68757df69d to 1 deployment-controller logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:04:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmaximumjackaltestesindexcle-1-66v5rh AddedInterface Add eth0 [10.129.2.24/23] from ovn-kubernetes multus logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:04:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmaximumjackaltestesindexcle-1-66v5rh.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:04:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmaximumjackaltestesindexcle-1-66v5rh.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:04:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmaximumjackaltestesindexcle-1-66v5rh.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:04:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmaximumjackaltestesindexcle-1-66v5rh.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:04:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmaximumjackaltestesindexcle-1-66v5rh.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:04:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmaximumjackaltestesindexcle-1-66v5rh.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:04:52 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmaximumjackaltestesindexcle-1-66v5rh.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:04:57 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmaximumjackaltestesindexcle-1-66v5rh.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:09 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-6d6b98d8b9-cl6ks Binding Scheduled Successfully assigned kuttl-test-maximum-jackal/test-es-index-cleaner-with-prefix-collector-6d6b98d8b9-cl6ks to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:09 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-collector-6d6b98d8b9 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-collector-6d6b98d8b9-cl6ks replicaset-controller logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:09 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-collector ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-collector-6d6b98d8b9 to 1 deployment-controller logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:09 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7dcd79bcdd-9rv42 Binding Scheduled Successfully assigned kuttl-test-maximum-jackal/test-es-index-cleaner-with-prefix-query-7dcd79bcdd-9rv42 to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:09 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-7dcd79bcdd SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-query-7dcd79bcdd-9rv42 replicaset-controller logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:09 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-query-7dcd79bcdd to 1 deployment-controller logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:10 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-6d6b98d8b9-cl6ks AddedInterface Add eth0 [10.131.0.59/23] from ovn-kubernetes multus logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:10 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-6d6b98d8b9-cl6ks.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:261a2c64c771323ad259e14802703bb790ee6b486b72c4939b487fbd83c30703" already present on machine kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:10 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-6d6b98d8b9-cl6ks.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:10 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-6d6b98d8b9-cl6ks.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:10 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7dcd79bcdd-9rv42 AddedInterface Add eth0 [10.128.2.48/23] from ovn-kubernetes multus logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:10 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7dcd79bcdd-9rv42.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:156ea63cbd4dcd8c44173f118dba74388adbf1e82493237a4c514bb14131d5b8" already present on machine kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:10 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7dcd79bcdd-9rv42.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:10 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7dcd79bcdd-9rv42.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:10 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7dcd79bcdd-9rv42.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:10 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7dcd79bcdd-9rv42.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:10 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7dcd79bcdd-9rv42.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:10 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7dcd79bcdd-9rv42.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" already present on machine kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:10 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7dcd79bcdd-9rv42.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:10 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7dcd79bcdd-9rv42.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:15 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7dcd79bcdd-9rv42.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:15 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7dcd79bcdd-9rv42.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:15 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-7dcd79bcdd-9rv42.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:15 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-7dcd79bcdd SuccessfulDelete Deleted pod: test-es-index-cleaner-with-prefix-query-7dcd79bcdd-9rv42 replicaset-controller logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:15 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled down replica set test-es-index-cleaner-with-prefix-query-7dcd79bcdd to 0 from 1 deployment-controller logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:16 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-679d7d4bf7-gb4cx Binding Scheduled Successfully assigned kuttl-test-maximum-jackal/test-es-index-cleaner-with-prefix-query-679d7d4bf7-gb4cx to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:16 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-679d7d4bf7-gb4cx AddedInterface Add eth0 [10.128.2.49/23] from ovn-kubernetes multus logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:16 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-679d7d4bf7 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-query-679d7d4bf7-gb4cx replicaset-controller logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:16 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-query-679d7d4bf7 to 1 deployment-controller logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:17 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-679d7d4bf7-gb4cx.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:156ea63cbd4dcd8c44173f118dba74388adbf1e82493237a4c514bb14131d5b8" already present on machine kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:17 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-679d7d4bf7-gb4cx.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:17 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-679d7d4bf7-gb4cx.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:17 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-679d7d4bf7-gb4cx.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:17 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-679d7d4bf7-gb4cx.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:17 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-679d7d4bf7-gb4cx.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:17 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-679d7d4bf7-gb4cx.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" already present on machine kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:17 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-679d7d4bf7-gb4cx.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:17 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-679d7d4bf7-gb4cx.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:21 +0000 UTC Normal Pod 00-report-span-6wl8r Binding Scheduled Successfully assigned kuttl-test-maximum-jackal/00-report-span-6wl8r to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:21 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-6wl8r job-controller logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:22 +0000 UTC Normal Pod 00-report-span-6wl8r AddedInterface Add eth0 [10.131.0.60/23] from ovn-kubernetes multus logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:22 +0000 UTC Normal Pod 00-report-span-6wl8r.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:22 +0000 UTC Normal Pod 00-report-span-6wl8r.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:22 +0000 UTC Normal Pod 00-report-span-6wl8r.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:24 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:24 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:24 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:54 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-6d6b98d8b9-cl6ks horizontal-pod-autoscaler logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:54 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-6d6b98d8b9-cl6ks horizontal-pod-autoscaler logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:05:54 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-6d6b98d8b9-cl6ks horizontal-pod-autoscaler logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:06:00 +0000 UTC Normal Job.batch 00-report-span Completed Job completed job-controller logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:07:00 +0000 UTC Normal Job.batch test-es-index-cleaner-with-prefix-es-index-cleaner-28603207 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-es-index-cleaner-2860320sdfpq job-controller logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:07:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2860320sdfpq Binding Scheduled Successfully assigned kuttl-test-maximum-jackal/test-es-index-cleaner-with-prefix-es-index-cleaner-2860320sdfpq to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:07:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2860320sdfpq AddedInterface Add eth0 [10.131.0.61/23] from ovn-kubernetes multus logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:07:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2860320sdfpq.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-index-cleaner-rhel8@sha256:58889396e0e99e83d7b7e003b9326dec39d020b0ddd57ac81f2eea81e3a39c8a" kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:07:00 +0000 UTC Normal CronJob.batch test-es-index-cleaner-with-prefix-es-index-cleaner SuccessfulCreate Created job test-es-index-cleaner-with-prefix-es-index-cleaner-28603207 cronjob-controller logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:07:02 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2860320sdfpq.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-index-cleaner-rhel8@sha256:58889396e0e99e83d7b7e003b9326dec39d020b0ddd57ac81f2eea81e3a39c8a" in 1.937s (1.937s including waiting) kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:07:02 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2860320sdfpq.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Created Created container test-es-index-cleaner-with-prefix-es-index-cleaner kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:07:02 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2860320sdfpq.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Started Started container test-es-index-cleaner-with-prefix-es-index-cleaner kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:07:05 +0000 UTC Normal Job.batch test-es-index-cleaner-with-prefix-es-index-cleaner-28603207 Completed Job completed job-controller logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:07:05 +0000 UTC Normal CronJob.batch test-es-index-cleaner-with-prefix-es-index-cleaner SawCompletedJob Saw completed job: test-es-index-cleaner-with-prefix-es-index-cleaner-28603207, status: Complete cronjob-controller logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:07:11 +0000 UTC Normal Pod 00-check-indices-th7mn Binding Scheduled Successfully assigned kuttl-test-maximum-jackal/00-check-indices-th7mn to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:07:11 +0000 UTC Normal Job.batch 00-check-indices SuccessfulCreate Created pod: 00-check-indices-th7mn job-controller logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:07:12 +0000 UTC Normal Pod 00-check-indices-th7mn AddedInterface Add eth0 [10.131.0.62/23] from ovn-kubernetes multus logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:07:12 +0000 UTC Normal Pod 00-check-indices-th7mn.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:07:12 +0000 UTC Normal Pod 00-check-indices-th7mn.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:07:12 +0000 UTC Normal Pod 00-check-indices-th7mn.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:07:15 | es-index-cleaner-autoprov | 2024-05-20 08:07:15 +0000 UTC Normal Job.batch 00-check-indices Completed Job completed job-controller logger.go:42: 08:07:15 | es-index-cleaner-autoprov | Deleting namespace: kuttl-test-maximum-jackal === CONT kuttl/harness/es-from-aio-to-production logger.go:42: 08:07:22 | es-from-aio-to-production | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:07:22 | es-from-aio-to-production | Creating namespace: kuttl-test-clever-amoeba logger.go:42: 08:07:22 | es-from-aio-to-production/0-install | starting test step 0-install logger.go:42: 08:07:22 | es-from-aio-to-production/0-install | Jaeger:kuttl-test-clever-amoeba/my-jaeger created logger.go:42: 08:07:28 | es-from-aio-to-production/0-install | test step completed 0-install logger.go:42: 08:07:28 | es-from-aio-to-production/1-smoke-test | starting test step 1-smoke-test logger.go:42: 08:07:28 | es-from-aio-to-production/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 08:07:30 | es-from-aio-to-production/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 08:07:37 | es-from-aio-to-production/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 08:07:37 | es-from-aio-to-production/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 08:07:38 | es-from-aio-to-production/1-smoke-test | job.batch/report-span created logger.go:42: 08:07:38 | es-from-aio-to-production/1-smoke-test | job.batch/check-span created logger.go:42: 08:07:50 | es-from-aio-to-production/1-smoke-test | test step completed 1-smoke-test logger.go:42: 08:07:50 | es-from-aio-to-production/3-install | starting test step 3-install logger.go:42: 08:07:50 | es-from-aio-to-production/3-install | Jaeger:kuttl-test-clever-amoeba/my-jaeger updated logger.go:42: 08:08:24 | es-from-aio-to-production/3-install | test step completed 3-install logger.go:42: 08:08:24 | es-from-aio-to-production/4-smoke-test | starting test step 4-smoke-test logger.go:42: 08:08:24 | es-from-aio-to-production/4-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 08:08:32 | es-from-aio-to-production/4-smoke-test | running command: [sh -c ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 08:08:33 | es-from-aio-to-production/4-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 08:08:33 | es-from-aio-to-production/4-smoke-test | job.batch/report-span unchanged logger.go:42: 08:08:33 | es-from-aio-to-production/4-smoke-test | job.batch/check-span unchanged logger.go:42: 08:08:33 | es-from-aio-to-production/4-smoke-test | test step completed 4-smoke-test logger.go:42: 08:08:33 | es-from-aio-to-production | es-from-aio-to-production events from ns kuttl-test-clever-amoeba: logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:26 +0000 UTC Normal Pod my-jaeger-678b45c64c-2bgtv Binding Scheduled Successfully assigned kuttl-test-clever-amoeba/my-jaeger-678b45c64c-2bgtv to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:26 +0000 UTC Normal ReplicaSet.apps my-jaeger-678b45c64c SuccessfulCreate Created pod: my-jaeger-678b45c64c-2bgtv replicaset-controller logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:26 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-678b45c64c to 1 deployment-controller logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:27 +0000 UTC Normal Pod my-jaeger-678b45c64c-2bgtv AddedInterface Add eth0 [10.129.2.25/23] from ovn-kubernetes multus logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:27 +0000 UTC Normal Pod my-jaeger-678b45c64c-2bgtv.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" already present on machine kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:27 +0000 UTC Normal Pod my-jaeger-678b45c64c-2bgtv.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:27 +0000 UTC Normal Pod my-jaeger-678b45c64c-2bgtv.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:27 +0000 UTC Normal Pod my-jaeger-678b45c64c-2bgtv.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:27 +0000 UTC Normal Pod my-jaeger-678b45c64c-2bgtv.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:27 +0000 UTC Normal Pod my-jaeger-678b45c64c-2bgtv.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:31 +0000 UTC Normal Pod my-jaeger-678b45c64c-2bgtv.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:31 +0000 UTC Normal Pod my-jaeger-678b45c64c-2bgtv.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:31 +0000 UTC Normal ReplicaSet.apps my-jaeger-678b45c64c SuccessfulDelete Deleted pod: my-jaeger-678b45c64c-2bgtv replicaset-controller logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:31 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-678b45c64c to 0 from 1 deployment-controller logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:32 +0000 UTC Normal Pod my-jaeger-6dd7645855-7fwnw Binding Scheduled Successfully assigned kuttl-test-clever-amoeba/my-jaeger-6dd7645855-7fwnw to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:32 +0000 UTC Normal ReplicaSet.apps my-jaeger-6dd7645855 SuccessfulCreate Created pod: my-jaeger-6dd7645855-7fwnw replicaset-controller logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:32 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-6dd7645855 to 1 deployment-controller logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:33 +0000 UTC Normal Pod my-jaeger-6dd7645855-7fwnw AddedInterface Add eth0 [10.129.2.26/23] from ovn-kubernetes multus logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:33 +0000 UTC Normal Pod my-jaeger-6dd7645855-7fwnw.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" already present on machine kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:33 +0000 UTC Normal Pod my-jaeger-6dd7645855-7fwnw.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:33 +0000 UTC Normal Pod my-jaeger-6dd7645855-7fwnw.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:33 +0000 UTC Normal Pod my-jaeger-6dd7645855-7fwnw.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:33 +0000 UTC Normal Pod my-jaeger-6dd7645855-7fwnw.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:33 +0000 UTC Normal Pod my-jaeger-6dd7645855-7fwnw.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:38 +0000 UTC Normal Pod check-span-htfsk Binding Scheduled Successfully assigned kuttl-test-clever-amoeba/check-span-htfsk to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:38 +0000 UTC Normal Pod check-span-htfsk AddedInterface Add eth0 [10.128.2.50/23] from ovn-kubernetes multus logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:38 +0000 UTC Normal Pod check-span-htfsk.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:38 +0000 UTC Normal Pod check-span-htfsk.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:38 +0000 UTC Normal Pod check-span-htfsk.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:38 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-htfsk job-controller logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:38 +0000 UTC Normal Pod report-span-bgn7v Binding Scheduled Successfully assigned kuttl-test-clever-amoeba/report-span-bgn7v to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:38 +0000 UTC Normal Pod report-span-bgn7v AddedInterface Add eth0 [10.131.0.63/23] from ovn-kubernetes multus logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:38 +0000 UTC Normal Pod report-span-bgn7v.spec.containers{report-span} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:38 +0000 UTC Normal Pod report-span-bgn7v.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:38 +0000 UTC Normal Pod report-span-bgn7v.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:38 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-bgn7v job-controller logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:49 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcleveramoebamyjaeger-1-66c894dc75vln Binding Scheduled Successfully assigned kuttl-test-clever-amoeba/elasticsearch-cdm-kuttltestcleveramoebamyjaeger-1-66c894dc75vln to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:54 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestcleveramoebamyjaeger-1-66c894dcbd SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestcleveramoebamyjaeger-1-66c894dc75vln replicaset-controller logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:54 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestcleveramoebamyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestcleveramoebamyjaeger-1-66c894dcbd to 1 deployment-controller logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcleveramoebamyjaeger-1-66c894dc75vln AddedInterface Add eth0 [10.131.0.64/23] from ovn-kubernetes multus logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcleveramoebamyjaeger-1-66c894dc75vln.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcleveramoebamyjaeger-1-66c894dc75vln.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcleveramoebamyjaeger-1-66c894dc75vln.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcleveramoebamyjaeger-1-66c894dc75vln.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcleveramoebamyjaeger-1-66c894dc75vln.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:07:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcleveramoebamyjaeger-1-66c894dc75vln.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:08:10 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestcleveramoebamyjaeger-1-66c894dc75vln.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:08:11 +0000 UTC Normal Job.batch report-span Completed Job completed job-controller logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:08:21 +0000 UTC Normal Pod my-jaeger-6dd7645855-7fwnw.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:08:21 +0000 UTC Normal Pod my-jaeger-6dd7645855-7fwnw.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:08:21 +0000 UTC Normal Pod my-jaeger-collector-76f849c475-zbh5t Binding Scheduled Successfully assigned kuttl-test-clever-amoeba/my-jaeger-collector-76f849c475-zbh5t to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:08:21 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-76f849c475 SuccessfulCreate Created pod: my-jaeger-collector-76f849c475-zbh5t replicaset-controller logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:08:21 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-76f849c475 to 1 deployment-controller logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:08:21 +0000 UTC Normal Pod my-jaeger-query-849c999c4b-6w9nm Binding Scheduled Successfully assigned kuttl-test-clever-amoeba/my-jaeger-query-849c999c4b-6w9nm to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:08:21 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-849c999c4b SuccessfulCreate Created pod: my-jaeger-query-849c999c4b-6w9nm replicaset-controller logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:08:21 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-849c999c4b to 1 deployment-controller logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:08:22 +0000 UTC Normal Pod my-jaeger-collector-76f849c475-zbh5t AddedInterface Add eth0 [10.128.2.51/23] from ovn-kubernetes multus logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:08:22 +0000 UTC Normal Pod my-jaeger-collector-76f849c475-zbh5t.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:261a2c64c771323ad259e14802703bb790ee6b486b72c4939b487fbd83c30703" already present on machine kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:08:22 +0000 UTC Normal Pod my-jaeger-collector-76f849c475-zbh5t.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:08:22 +0000 UTC Normal Pod my-jaeger-collector-76f849c475-zbh5t.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:08:22 +0000 UTC Normal Pod my-jaeger-query-849c999c4b-6w9nm AddedInterface Add eth0 [10.128.2.52/23] from ovn-kubernetes multus logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:08:22 +0000 UTC Normal Pod my-jaeger-query-849c999c4b-6w9nm.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:156ea63cbd4dcd8c44173f118dba74388adbf1e82493237a4c514bb14131d5b8" already present on machine kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:08:22 +0000 UTC Normal Pod my-jaeger-query-849c999c4b-6w9nm.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:08:22 +0000 UTC Normal Pod my-jaeger-query-849c999c4b-6w9nm.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:08:22 +0000 UTC Normal Pod my-jaeger-query-849c999c4b-6w9nm.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:08:22 +0000 UTC Normal Pod my-jaeger-query-849c999c4b-6w9nm.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:08:22 +0000 UTC Normal Pod my-jaeger-query-849c999c4b-6w9nm.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:08:22 +0000 UTC Normal Pod my-jaeger-query-849c999c4b-6w9nm.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" already present on machine kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:08:22 +0000 UTC Normal Pod my-jaeger-query-849c999c4b-6w9nm.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | 2024-05-20 08:08:22 +0000 UTC Normal Pod my-jaeger-query-849c999c4b-6w9nm.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:08:33 | es-from-aio-to-production | Deleting namespace: kuttl-test-clever-amoeba === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (711.09s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.91s) --- PASS: kuttl/harness/es-multiinstance (93.91s) --- PASS: kuttl/harness/es-increasing-replicas (113.70s) --- PASS: kuttl/harness/es-rollover-autoprov (245.72s) --- PASS: kuttl/harness/es-simple-prod (5.94s) --- PASS: kuttl/harness/es-index-cleaner-autoprov (168.29s) --- PASS: kuttl/harness/es-from-aio-to-production (77.45s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name elasticsearch --report --output /logs/artifacts/elasticsearch.xml ./artifacts/kuttl-report.xml time="2024-05-20T08:08:40Z" level=debug msg="Setting a new name for the test suites" time="2024-05-20T08:08:40Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-05-20T08:08:40Z" level=debug msg="normalizing test case names" time="2024-05-20T08:08:40Z" level=debug msg="elasticsearch/artifacts -> elasticsearch_artifacts" time="2024-05-20T08:08:40Z" level=debug msg="elasticsearch/es-multiinstance -> elasticsearch_es_multiinstance" time="2024-05-20T08:08:40Z" level=debug msg="elasticsearch/es-increasing-replicas -> elasticsearch_es_increasing_replicas" time="2024-05-20T08:08:40Z" level=debug msg="elasticsearch/es-rollover-autoprov -> elasticsearch_es_rollover_autoprov" time="2024-05-20T08:08:40Z" level=debug msg="elasticsearch/es-simple-prod -> elasticsearch_es_simple_prod" time="2024-05-20T08:08:40Z" level=debug msg="elasticsearch/es-index-cleaner-autoprov -> elasticsearch_es_index_cleaner_autoprov" time="2024-05-20T08:08:40Z" level=debug msg="elasticsearch/es-from-aio-to-production -> elasticsearch_es_from_aio_to_production" +-----------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------+--------+ | elasticsearch_artifacts | passed | | elasticsearch_es_multiinstance | passed | | elasticsearch_es_increasing_replicas | passed | | elasticsearch_es_rollover_autoprov | passed | | elasticsearch_es_simple_prod | passed | | elasticsearch_es_index_cleaner_autoprov | passed | | elasticsearch_es_from_aio_to_production | passed | +-----------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh examples false true + '[' 3 -ne 3 ']' + test_suite_name=examples + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/examples.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-examples make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ VERTX_IMG=jaegertracing/vertx-create-span:operator-e2e-tests \ ./tests/e2e/examples/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-05-19-083311 True False 26m Cluster version is 4.16.0-0.nightly-2024-05-19-083311' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-05-19-083311 True False 26m Cluster version is 4.16.0-0.nightly-2024-05-19-083311' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/examples/render.sh ++ export SUITE_DIR=./tests/e2e/examples ++ SUITE_DIR=./tests/e2e/examples ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/examples ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test examples-agent-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-agent-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-with-priority-class\e[0m' Rendering files for test examples-agent-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + mkdir -p examples-agent-with-priority-class + cd examples-agent-with-priority-class + example_name=agent-with-priority-class + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + '[' true '!=' true ']' + render_install_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + install_file=./02-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/agent-with-priority-class.yaml -o ./02-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./02-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./02-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./02-install.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + JAEGER_NAME=agent-as-daemonset + local jaeger_strategy ++ get_jaeger_strategy ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./02-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./02-install.yaml ++ strategy=DaemonSet ++ '[' DaemonSet = null ']' ++ echo DaemonSet ++ return 0 + jaeger_strategy=DaemonSet + '[' DaemonSet = DaemonSet ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./02-assert.yaml + render_smoke_test_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + jaeger_name=agent-as-daemonset + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test agent-as-daemonset true 02 + '[' 3 -ne 3 ']' + jaeger=agent-as-daemonset + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + export JAEGER_NAME=agent-as-daemonset + JAEGER_NAME=agent-as-daemonset + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-all-in-one-with-options + '[' 1 -ne 1 ']' + test_name=examples-all-in-one-with-options + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-all-in-one-with-options' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-all-in-one-with-options\e[0m' Rendering files for test examples-all-in-one-with-options + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-with-priority-class + '[' examples-agent-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-all-in-one-with-options + cd examples-all-in-one-with-options + example_name=all-in-one-with-options + render_install_example all-in-one-with-options 00 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/all-in-one-with-options.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + JAEGER_NAME=my-jaeger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.metadata.name="my-jaeger"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i 'del(.spec.allInOne.image)' ./00-install.yaml + render_smoke_test_example all-in-one-with-options 01 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + jaeger_name=my-jaeger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + sed -i s~my-jaeger-query:443~my-jaeger-query:443/jaeger~gi ./01-smoke-test.yaml + start_test examples-business-application-injected-sidecar + '[' 1 -ne 1 ']' + test_name=examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-business-application-injected-sidecar' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-business-application-injected-sidecar\e[0m' Rendering files for test examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-all-in-one-with-options + '[' examples-all-in-one-with-options '!=' _build ']' + cd .. + mkdir -p examples-business-application-injected-sidecar + cd examples-business-application-injected-sidecar + example_name=simplest + cp /tmp/jaeger-tests/examples/business-application-injected-sidecar.yaml ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].image=strenv(VERTX_IMG)' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.path="/"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.port=8080' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.initialDelaySeconds=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.failureThreshold=3' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.periodSeconds=10' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.successThreshold=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.timeoutSeconds=1' ./00-install.yaml + render_install_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example simplest 02 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 02 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-collector-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-collector-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-collector-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-collector-with-priority-class\e[0m' Rendering files for test examples-collector-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-business-application-injected-sidecar + '[' examples-business-application-injected-sidecar '!=' _build ']' + cd .. + mkdir -p examples-collector-with-priority-class + cd examples-collector-with-priority-class + example_name=collector-with-priority-class + render_install_example collector-with-priority-class 00 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/collector-with-priority-class.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + JAEGER_NAME=collector-with-high-priority + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example collector-with-priority-class 01 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + jaeger_name=collector-with-high-priority + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test collector-with-high-priority true 01 + '[' 3 -ne 3 ']' + jaeger=collector-with-high-priority + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + export JAEGER_NAME=collector-with-high-priority + JAEGER_NAME=collector-with-high-priority + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-service-types + '[' 1 -ne 1 ']' + test_name=examples-service-types + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-service-types' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-service-types\e[0m' Rendering files for test examples-service-types + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-collector-with-priority-class + '[' examples-collector-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-service-types + cd examples-service-types + example_name=service-types + render_install_example service-types 00 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/service-types.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + JAEGER_NAME=service-types + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example service-types 01 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/service-types.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/service-types.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/service-types.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/service-types.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + jaeger_name=service-types + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test service-types true 01 + '[' 3 -ne 3 ']' + jaeger=service-types + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + export JAEGER_NAME=service-types + JAEGER_NAME=service-types + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod + '[' 1 -ne 1 ']' + test_name=examples-simple-prod + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod\e[0m' Rendering files for test examples-simple-prod + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-service-types + '[' examples-service-types '!=' _build ']' + cd .. + mkdir -p examples-simple-prod + cd examples-simple-prod + example_name=simple-prod + render_install_example simple-prod 01 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod 02 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod-with-volumes + '[' 1 -ne 1 ']' + test_name=examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod-with-volumes' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod-with-volumes\e[0m' Rendering files for test examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod + '[' examples-simple-prod '!=' _build ']' + cd .. + mkdir -p examples-simple-prod-with-volumes + cd examples-simple-prod-with-volumes + example_name=simple-prod-with-volumes + render_install_example simple-prod-with-volumes 01 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod-with-volumes 02 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + /tmp/jaeger-tests/bin/gomplate -f ./03-check-volume.yaml.template -o 03-check-volume.yaml + start_test examples-simplest + '[' 1 -ne 1 ']' + test_name=examples-simplest + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simplest' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simplest\e[0m' Rendering files for test examples-simplest + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod-with-volumes + '[' examples-simple-prod-with-volumes '!=' _build ']' + cd .. + mkdir -p examples-simplest + cd examples-simplest + example_name=simplest + render_install_example simplest 00 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 01 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger + '[' 1 -ne 1 ']' + test_name=examples-with-badger + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger\e[0m' Rendering files for test examples-with-badger + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simplest + '[' examples-simplest '!=' _build ']' + cd .. + mkdir -p examples-with-badger + cd examples-with-badger + example_name=with-badger + render_install_example with-badger 00 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + JAEGER_NAME=with-badger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger 01 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + jaeger_name=with-badger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + export JAEGER_NAME=with-badger + JAEGER_NAME=with-badger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger-and-volume + '[' 1 -ne 1 ']' + test_name=examples-with-badger-and-volume + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger-and-volume' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger-and-volume\e[0m' Rendering files for test examples-with-badger-and-volume + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger + '[' examples-with-badger '!=' _build ']' + cd .. + mkdir -p examples-with-badger-and-volume + cd examples-with-badger-and-volume + example_name=with-badger-and-volume + render_install_example with-badger-and-volume 00 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger-and-volume.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + JAEGER_NAME=with-badger-and-volume + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger-and-volume 01 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + jaeger_name=with-badger-and-volume + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger-and-volume true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger-and-volume + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + export JAEGER_NAME=with-badger-and-volume + JAEGER_NAME=with-badger-and-volume + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-cassandra + '[' 1 -ne 1 ']' + test_name=examples-with-cassandra + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-cassandra' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-cassandra\e[0m' Rendering files for test examples-with-cassandra + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger-and-volume + '[' examples-with-badger-and-volume '!=' _build ']' + cd .. + mkdir -p examples-with-cassandra + cd examples-with-cassandra + example_name=with-cassandra + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-cassandra 01 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-cassandra.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + JAEGER_NAME=with-cassandra + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-cassandra 02 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-cassandra.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-cassandra.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + jaeger_name=with-cassandra + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-cassandra true 02 + '[' 3 -ne 3 ']' + jaeger=with-cassandra + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + export JAEGER_NAME=with-cassandra + JAEGER_NAME=with-cassandra + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-sampling + '[' 1 -ne 1 ']' + test_name=examples-with-sampling + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-sampling' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-sampling\e[0m' Rendering files for test examples-with-sampling + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-cassandra + '[' examples-with-cassandra '!=' _build ']' + cd .. + mkdir -p examples-with-sampling + cd examples-with-sampling + export example_name=with-sampling + example_name=with-sampling + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-sampling 01 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-sampling.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + JAEGER_NAME=with-sampling + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-sampling 02 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-sampling.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-sampling.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + jaeger_name=with-sampling + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-sampling true 02 + '[' 3 -ne 3 ']' + jaeger=with-sampling + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + export JAEGER_NAME=with-sampling + JAEGER_NAME=with-sampling + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-agent-as-daemonset + '[' 1 -ne 1 ']' + test_name=examples-agent-as-daemonset + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-as-daemonset' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-as-daemonset\e[0m' Rendering files for test examples-agent-as-daemonset + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-sampling + '[' examples-with-sampling '!=' _build ']' + cd .. + mkdir -p examples-agent-as-daemonset + cd examples-agent-as-daemonset + '[' true = true ']' + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/agent-as-daemonset.yaml -o 02-install.yaml + '[' true = true ']' + start_test examples-openshift-with-htpasswd + '[' 1 -ne 1 ']' + test_name=examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-openshift-with-htpasswd' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-openshift-with-htpasswd\e[0m' Rendering files for test examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-as-daemonset + '[' examples-agent-as-daemonset '!=' _build ']' + cd .. + mkdir -p examples-openshift-with-htpasswd + cd examples-openshift-with-htpasswd + export JAEGER_NAME=with-htpasswd + JAEGER_NAME=with-htpasswd + export JAEGER_USERNAME=awesomeuser + JAEGER_USERNAME=awesomeuser + export JAEGER_PASSWORD=awesomepassword + JAEGER_PASSWORD=awesomepassword + export 'JAEGER_USER_PASSWORD_HASH=awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' + JAEGER_USER_PASSWORD_HASH='awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ echo 'awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ base64 + SECRET=YXdlc29tZXVzZXI6e1NIQX11VWRxUFZVeXFOQm1FUlUwUXhqM0tGYVpuanc9Cg== + /tmp/jaeger-tests/bin/gomplate -f ./00-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/with-htpasswd.yaml -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + export 'GET_URL_COMMAND=kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + export 'URL=https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + INSECURE=true + JAEGER_USERNAME= + JAEGER_PASSWORD= + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-unsecured.yaml + JAEGER_USERNAME=wronguser + JAEGER_PASSWORD=wrongpassword + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-check-unauthorized.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./04-check-authorized.yaml + skip_test examples-agent-as-daemonset 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-agent-as-daemonset + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-openshift-with-htpasswd + '[' examples-openshift-with-htpasswd '!=' _build ']' + cd .. + rm -rf examples-agent-as-daemonset + warning 'examples-agent-as-daemonset: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-agent-as-daemonset: This test is flaky in Prow CI\e[0m' WAR: examples-agent-as-daemonset: This test is flaky in Prow CI + skip_test examples-with-badger-and-volume 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-with-badger-and-volume + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + rm -rf examples-with-badger-and-volume + warning 'examples-with-badger-and-volume: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-with-badger-and-volume: This test is flaky in Prow CI\e[0m' WAR: examples-with-badger-and-volume: This test is flaky in Prow CI + skip_test examples-collector-with-priority-class 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-collector-with-priority-class + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + rm -rf examples-collector-with-priority-class + warning 'examples-collector-with-priority-class: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-collector-with-priority-class: This test is flaky in Prow CI\e[0m' WAR: examples-collector-with-priority-class: This test is flaky in Prow CI make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running examples E2E tests' Running examples E2E tests + cd tests/e2e/examples/_build + set +e + KUBECONFIG=/tmp/kubeconfig-559878434 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 12 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/examples-agent-with-priority-class === PAUSE kuttl/harness/examples-agent-with-priority-class === RUN kuttl/harness/examples-all-in-one-with-options === PAUSE kuttl/harness/examples-all-in-one-with-options === RUN kuttl/harness/examples-business-application-injected-sidecar === PAUSE kuttl/harness/examples-business-application-injected-sidecar === RUN kuttl/harness/examples-openshift-with-htpasswd === PAUSE kuttl/harness/examples-openshift-with-htpasswd === RUN kuttl/harness/examples-service-types === PAUSE kuttl/harness/examples-service-types === RUN kuttl/harness/examples-simple-prod === PAUSE kuttl/harness/examples-simple-prod === RUN kuttl/harness/examples-simple-prod-with-volumes === PAUSE kuttl/harness/examples-simple-prod-with-volumes === RUN kuttl/harness/examples-simplest === PAUSE kuttl/harness/examples-simplest === RUN kuttl/harness/examples-with-badger === PAUSE kuttl/harness/examples-with-badger === RUN kuttl/harness/examples-with-cassandra === PAUSE kuttl/harness/examples-with-cassandra === RUN kuttl/harness/examples-with-sampling === PAUSE kuttl/harness/examples-with-sampling === CONT kuttl/harness/artifacts logger.go:42: 08:09:11 | artifacts | Creating namespace: kuttl-test-renewed-flamingo logger.go:42: 08:09:11 | artifacts | artifacts events from ns kuttl-test-renewed-flamingo: logger.go:42: 08:09:11 | artifacts | Deleting namespace: kuttl-test-renewed-flamingo === CONT kuttl/harness/examples-simple-prod logger.go:42: 08:09:17 | examples-simple-prod | Creating namespace: kuttl-test-up-cow logger.go:42: 08:09:17 | examples-simple-prod/1-install | starting test step 1-install logger.go:42: 08:09:17 | examples-simple-prod/1-install | Jaeger:kuttl-test-up-cow/simple-prod created logger.go:42: 08:09:56 | examples-simple-prod/1-install | test step completed 1-install logger.go:42: 08:09:56 | examples-simple-prod/2-smoke-test | starting test step 2-smoke-test logger.go:42: 08:09:56 | examples-simple-prod/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 08:09:57 | examples-simple-prod/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 08:10:04 | examples-simple-prod/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337 JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 08:10:04 | examples-simple-prod/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 08:10:05 | examples-simple-prod/2-smoke-test | job.batch/report-span created logger.go:42: 08:10:05 | examples-simple-prod/2-smoke-test | job.batch/check-span created logger.go:42: 08:10:17 | examples-simple-prod/2-smoke-test | test step completed 2-smoke-test logger.go:42: 08:10:17 | examples-simple-prod | examples-simple-prod events from ns kuttl-test-up-cow: logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestupcowsimpleprod-1-6cdb88fc44-twmdp Binding Scheduled Successfully assigned kuttl-test-up-cow/elasticsearch-cdm-kuttltestupcowsimpleprod-1-6cdb88fc44-twmdp to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:24 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestupcowsimpleprod-1-6cdb88fc44 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestupcowsimpleprod-1-6cdb88fc44-twmdp replicaset-controller logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:24 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestupcowsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestupcowsimpleprod-1-6cdb88fc44 to 1 deployment-controller logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestupcowsimpleprod-1-6cdb88fc44-twmdp AddedInterface Add eth0 [10.129.2.27/23] from ovn-kubernetes multus logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestupcowsimpleprod-1-6cdb88fc44-twmdp.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestupcowsimpleprod-1-6cdb88fc44-twmdp.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestupcowsimpleprod-1-6cdb88fc44-twmdp.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestupcowsimpleprod-1-6cdb88fc44-twmdp.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestupcowsimpleprod-1-6cdb88fc44-twmdp.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestupcowsimpleprod-1-6cdb88fc44-twmdp.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:35 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestupcowsimpleprod-1-6cdb88fc44-twmdp.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:40 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestupcowsimpleprod-1-6cdb88fc44-twmdp.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:51 +0000 UTC Normal Pod simple-prod-collector-859bd4847c-6z59j Binding Scheduled Successfully assigned kuttl-test-up-cow/simple-prod-collector-859bd4847c-6z59j to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:51 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-859bd4847c SuccessfulCreate Created pod: simple-prod-collector-859bd4847c-6z59j replicaset-controller logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:51 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-859bd4847c to 1 deployment-controller logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:51 +0000 UTC Normal Pod simple-prod-query-548d66649b-d8ftk Binding Scheduled Successfully assigned kuttl-test-up-cow/simple-prod-query-548d66649b-d8ftk to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:51 +0000 UTC Normal ReplicaSet.apps simple-prod-query-548d66649b SuccessfulCreate Created pod: simple-prod-query-548d66649b-d8ftk replicaset-controller logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:51 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-548d66649b to 1 deployment-controller logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:52 +0000 UTC Warning Pod simple-prod-collector-859bd4847c-6z59j FailedMount MountVolume.SetUp failed for volume "kube-api-access-xg7jb" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:52 +0000 UTC Normal Pod simple-prod-query-548d66649b-d8ftk AddedInterface Add eth0 [10.128.2.53/23] from ovn-kubernetes multus logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:52 +0000 UTC Normal Pod simple-prod-query-548d66649b-d8ftk.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:156ea63cbd4dcd8c44173f118dba74388adbf1e82493237a4c514bb14131d5b8" already present on machine kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:52 +0000 UTC Normal Pod simple-prod-query-548d66649b-d8ftk.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:52 +0000 UTC Normal Pod simple-prod-query-548d66649b-d8ftk.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:52 +0000 UTC Normal Pod simple-prod-query-548d66649b-d8ftk.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:52 +0000 UTC Normal Pod simple-prod-query-548d66649b-d8ftk.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:52 +0000 UTC Normal Pod simple-prod-query-548d66649b-d8ftk.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:52 +0000 UTC Normal Pod simple-prod-query-548d66649b-d8ftk.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" already present on machine kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:52 +0000 UTC Normal Pod simple-prod-query-548d66649b-d8ftk.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:52 +0000 UTC Normal Pod simple-prod-query-548d66649b-d8ftk.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:53 +0000 UTC Normal Pod simple-prod-collector-859bd4847c-6z59j AddedInterface Add eth0 [10.131.0.65/23] from ovn-kubernetes multus logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:53 +0000 UTC Normal Pod simple-prod-collector-859bd4847c-6z59j.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:261a2c64c771323ad259e14802703bb790ee6b486b72c4939b487fbd83c30703" already present on machine kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:54 +0000 UTC Normal Pod simple-prod-collector-859bd4847c-6z59j.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:09:54 +0000 UTC Normal Pod simple-prod-collector-859bd4847c-6z59j.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:00 +0000 UTC Normal Pod simple-prod-query-548d66649b-d8ftk.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:00 +0000 UTC Normal Pod simple-prod-query-548d66649b-d8ftk.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:00 +0000 UTC Normal Pod simple-prod-query-548d66649b-d8ftk.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:00 +0000 UTC Normal ReplicaSet.apps simple-prod-query-548d66649b SuccessfulDelete Deleted pod: simple-prod-query-548d66649b-d8ftk replicaset-controller logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:00 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-548d66649b to 0 from 1 deployment-controller logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:01 +0000 UTC Normal Pod simple-prod-query-66744ffd8c-pb6pj Binding Scheduled Successfully assigned kuttl-test-up-cow/simple-prod-query-66744ffd8c-pb6pj to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:01 +0000 UTC Normal Pod simple-prod-query-66744ffd8c-pb6pj AddedInterface Add eth0 [10.128.2.54/23] from ovn-kubernetes multus logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:01 +0000 UTC Normal Pod simple-prod-query-66744ffd8c-pb6pj.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:156ea63cbd4dcd8c44173f118dba74388adbf1e82493237a4c514bb14131d5b8" already present on machine kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:01 +0000 UTC Normal Pod simple-prod-query-66744ffd8c-pb6pj.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:01 +0000 UTC Normal Pod simple-prod-query-66744ffd8c-pb6pj.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:01 +0000 UTC Normal Pod simple-prod-query-66744ffd8c-pb6pj.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:01 +0000 UTC Normal Pod simple-prod-query-66744ffd8c-pb6pj.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:01 +0000 UTC Normal Pod simple-prod-query-66744ffd8c-pb6pj.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:01 +0000 UTC Normal ReplicaSet.apps simple-prod-query-66744ffd8c SuccessfulCreate Created pod: simple-prod-query-66744ffd8c-pb6pj replicaset-controller logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:01 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-66744ffd8c to 1 deployment-controller logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:02 +0000 UTC Normal Pod simple-prod-query-66744ffd8c-pb6pj.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" already present on machine kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:02 +0000 UTC Normal Pod simple-prod-query-66744ffd8c-pb6pj.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:02 +0000 UTC Normal Pod simple-prod-query-66744ffd8c-pb6pj.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:05 +0000 UTC Normal Pod check-span-q44cm Binding Scheduled Successfully assigned kuttl-test-up-cow/check-span-q44cm to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:05 +0000 UTC Normal Pod check-span-q44cm AddedInterface Add eth0 [10.131.0.67/23] from ovn-kubernetes multus logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:05 +0000 UTC Normal Pod check-span-q44cm.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:05 +0000 UTC Normal Pod check-span-q44cm.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:05 +0000 UTC Normal Pod check-span-q44cm.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:05 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-q44cm job-controller logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:05 +0000 UTC Normal Pod report-span-4xwfl Binding Scheduled Successfully assigned kuttl-test-up-cow/report-span-4xwfl to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:05 +0000 UTC Normal Pod report-span-4xwfl AddedInterface Add eth0 [10.131.0.66/23] from ovn-kubernetes multus logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:05 +0000 UTC Normal Pod report-span-4xwfl.spec.containers{report-span} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:05 +0000 UTC Normal Pod report-span-4xwfl.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:05 +0000 UTC Normal Pod report-span-4xwfl.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:05 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-4xwfl job-controller logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:06 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:06 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:06 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:10:17 | examples-simple-prod | 2024-05-20 08:10:16 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 08:10:17 | examples-simple-prod | Deleting namespace: kuttl-test-up-cow === CONT kuttl/harness/examples-with-sampling logger.go:42: 08:10:29 | examples-with-sampling | Creating namespace: kuttl-test-dashing-hippo logger.go:42: 08:10:29 | examples-with-sampling/0-install | starting test step 0-install logger.go:42: 08:10:29 | examples-with-sampling/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 08:10:29 | examples-with-sampling/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 08:10:29 | examples-with-sampling/0-install | >>>> Creating namespace kuttl-test-dashing-hippo logger.go:42: 08:10:29 | examples-with-sampling/0-install | kubectl create namespace kuttl-test-dashing-hippo 2>&1 | grep -v "already exists" || true logger.go:42: 08:10:30 | examples-with-sampling/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-dashing-hippo 2>&1 | grep -v "already exists" || true logger.go:42: 08:10:30 | examples-with-sampling/0-install | service/cassandra created logger.go:42: 08:10:30 | examples-with-sampling/0-install | statefulset.apps/cassandra created logger.go:42: 08:10:30 | examples-with-sampling/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 08:10:41 | examples-with-sampling/0-install | test step completed 0-install logger.go:42: 08:10:41 | examples-with-sampling/1-install | starting test step 1-install logger.go:42: 08:10:41 | examples-with-sampling/1-install | Jaeger:kuttl-test-dashing-hippo/with-sampling created logger.go:42: 08:10:48 | examples-with-sampling/1-install | test step completed 1-install logger.go:42: 08:10:48 | examples-with-sampling/2-smoke-test | starting test step 2-smoke-test logger.go:42: 08:10:48 | examples-with-sampling/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-sampling /dev/null] logger.go:42: 08:10:49 | examples-with-sampling/2-smoke-test | Warning: resource jaegers/with-sampling is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 08:10:56 | examples-with-sampling/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337 JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 08:10:56 | examples-with-sampling/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 08:10:57 | examples-with-sampling/2-smoke-test | job.batch/report-span created logger.go:42: 08:10:57 | examples-with-sampling/2-smoke-test | job.batch/check-span created logger.go:42: 08:11:09 | examples-with-sampling/2-smoke-test | test step completed 2-smoke-test logger.go:42: 08:11:09 | examples-with-sampling/3- | starting test step 3- logger.go:42: 08:11:09 | examples-with-sampling/3- | test step completed 3- logger.go:42: 08:11:09 | examples-with-sampling | examples-with-sampling events from ns kuttl-test-dashing-hippo: logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:30 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-dashing-hippo/cassandra-0 to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:30 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:31 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.131.0.68/23] from ovn-kubernetes multus logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:31 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:35 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 4.276s (4.276s including waiting) kubelet logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:35 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:35 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:35 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-dashing-hippo/cassandra-1 to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:35 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:36 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.129.2.28/23] from ovn-kubernetes multus logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:36 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:40 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 4.247s (4.247s including waiting) kubelet logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:40 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:40 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:45 +0000 UTC Normal Pod with-sampling-86b8bf764c-rzftc Binding Scheduled Successfully assigned kuttl-test-dashing-hippo/with-sampling-86b8bf764c-rzftc to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:45 +0000 UTC Normal ReplicaSet.apps with-sampling-86b8bf764c SuccessfulCreate Created pod: with-sampling-86b8bf764c-rzftc replicaset-controller logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:45 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-86b8bf764c to 1 deployment-controller logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:46 +0000 UTC Normal Pod with-sampling-86b8bf764c-rzftc AddedInterface Add eth0 [10.129.2.29/23] from ovn-kubernetes multus logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:46 +0000 UTC Normal Pod with-sampling-86b8bf764c-rzftc.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" already present on machine kubelet logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:46 +0000 UTC Normal Pod with-sampling-86b8bf764c-rzftc.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:46 +0000 UTC Normal Pod with-sampling-86b8bf764c-rzftc.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:46 +0000 UTC Normal Pod with-sampling-86b8bf764c-rzftc.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:46 +0000 UTC Normal Pod with-sampling-86b8bf764c-rzftc.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:46 +0000 UTC Normal Pod with-sampling-86b8bf764c-rzftc.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:52 +0000 UTC Normal Pod with-sampling-66b648fd6f-f5skg Binding Scheduled Successfully assigned kuttl-test-dashing-hippo/with-sampling-66b648fd6f-f5skg to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:52 +0000 UTC Normal ReplicaSet.apps with-sampling-66b648fd6f SuccessfulCreate Created pod: with-sampling-66b648fd6f-f5skg replicaset-controller logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:52 +0000 UTC Normal Pod with-sampling-86b8bf764c-rzftc.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:52 +0000 UTC Normal Pod with-sampling-86b8bf764c-rzftc.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:52 +0000 UTC Normal ReplicaSet.apps with-sampling-86b8bf764c SuccessfulDelete Deleted pod: with-sampling-86b8bf764c-rzftc replicaset-controller logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:52 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled down replica set with-sampling-86b8bf764c to 0 from 1 deployment-controller logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:52 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-66b648fd6f to 1 deployment-controller logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:53 +0000 UTC Normal Pod with-sampling-66b648fd6f-f5skg AddedInterface Add eth0 [10.129.2.30/23] from ovn-kubernetes multus logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:53 +0000 UTC Normal Pod with-sampling-66b648fd6f-f5skg.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" already present on machine kubelet logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:53 +0000 UTC Normal Pod with-sampling-66b648fd6f-f5skg.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:53 +0000 UTC Normal Pod with-sampling-66b648fd6f-f5skg.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:53 +0000 UTC Normal Pod with-sampling-66b648fd6f-f5skg.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:53 +0000 UTC Normal Pod with-sampling-66b648fd6f-f5skg.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:53 +0000 UTC Normal Pod with-sampling-66b648fd6f-f5skg.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:57 +0000 UTC Normal Pod check-span-dn4bl Binding Scheduled Successfully assigned kuttl-test-dashing-hippo/check-span-dn4bl to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:57 +0000 UTC Normal Pod check-span-dn4bl AddedInterface Add eth0 [10.131.0.69/23] from ovn-kubernetes multus logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:57 +0000 UTC Normal Pod check-span-dn4bl.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:57 +0000 UTC Normal Pod check-span-dn4bl.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:57 +0000 UTC Normal Pod check-span-dn4bl.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:57 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-dn4bl job-controller logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:57 +0000 UTC Normal Pod report-span-rdqvq Binding Scheduled Successfully assigned kuttl-test-dashing-hippo/report-span-rdqvq to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:57 +0000 UTC Normal Pod report-span-rdqvq AddedInterface Add eth0 [10.128.2.55/23] from ovn-kubernetes multus logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:57 +0000 UTC Normal Pod report-span-rdqvq.spec.containers{report-span} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:57 +0000 UTC Normal Pod report-span-rdqvq.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:57 +0000 UTC Normal Pod report-span-rdqvq.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:10:57 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-rdqvq job-controller logger.go:42: 08:11:09 | examples-with-sampling | 2024-05-20 08:11:08 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 08:11:09 | examples-with-sampling | Deleting namespace: kuttl-test-dashing-hippo === CONT kuttl/harness/examples-with-cassandra logger.go:42: 08:11:27 | examples-with-cassandra | Creating namespace: kuttl-test-prepared-finch logger.go:42: 08:11:27 | examples-with-cassandra/0-install | starting test step 0-install logger.go:42: 08:11:27 | examples-with-cassandra/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 08:11:27 | examples-with-cassandra/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 08:11:27 | examples-with-cassandra/0-install | >>>> Creating namespace kuttl-test-prepared-finch logger.go:42: 08:11:27 | examples-with-cassandra/0-install | kubectl create namespace kuttl-test-prepared-finch 2>&1 | grep -v "already exists" || true logger.go:42: 08:11:27 | examples-with-cassandra/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-prepared-finch 2>&1 | grep -v "already exists" || true logger.go:42: 08:11:28 | examples-with-cassandra/0-install | service/cassandra created logger.go:42: 08:11:28 | examples-with-cassandra/0-install | statefulset.apps/cassandra created logger.go:42: 08:11:28 | examples-with-cassandra/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 08:11:31 | examples-with-cassandra/0-install | test step completed 0-install logger.go:42: 08:11:31 | examples-with-cassandra/1-install | starting test step 1-install logger.go:42: 08:11:31 | examples-with-cassandra/1-install | Jaeger:kuttl-test-prepared-finch/with-cassandra created logger.go:42: 08:11:50 | examples-with-cassandra/1-install | test step completed 1-install logger.go:42: 08:11:50 | examples-with-cassandra/2-smoke-test | starting test step 2-smoke-test logger.go:42: 08:11:50 | examples-with-cassandra/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-cassandra /dev/null] logger.go:42: 08:11:52 | examples-with-cassandra/2-smoke-test | Warning: resource jaegers/with-cassandra is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 08:12:39 | examples-with-cassandra/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337 JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 08:12:39 | examples-with-cassandra/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 08:12:40 | examples-with-cassandra/2-smoke-test | job.batch/report-span created logger.go:42: 08:12:40 | examples-with-cassandra/2-smoke-test | job.batch/check-span created logger.go:42: 08:12:51 | examples-with-cassandra/2-smoke-test | test step completed 2-smoke-test logger.go:42: 08:12:51 | examples-with-cassandra | examples-with-cassandra events from ns kuttl-test-prepared-finch: logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:28 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-prepared-finch/cassandra-0 to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:28 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.129.2.31/23] from ovn-kubernetes multus logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:28 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:29 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:29 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:29 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:29 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-prepared-finch/cassandra-1 to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:29 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:30 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.131.0.70/23] from ovn-kubernetes multus logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:30 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:30 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:30 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:34 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-vrg6z Binding Scheduled Successfully assigned kuttl-test-prepared-finch/with-cassandra-cassandra-schema-job-vrg6z to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:34 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job SuccessfulCreate Created pod: with-cassandra-cassandra-schema-job-vrg6z job-controller logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:35 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-vrg6z AddedInterface Add eth0 [10.128.2.56/23] from ovn-kubernetes multus logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:35 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-vrg6z.spec.containers{with-cassandra-cassandra-schema-job} Pulling Pulling image "jaegertracing/jaeger-cassandra-schema:1.53.0" kubelet logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:40 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-vrg6z.spec.containers{with-cassandra-cassandra-schema-job} Pulled Successfully pulled image "jaegertracing/jaeger-cassandra-schema:1.53.0" in 5.126s (5.126s including waiting) kubelet logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:40 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-vrg6z.spec.containers{with-cassandra-cassandra-schema-job} Created Created container with-cassandra-cassandra-schema-job kubelet logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:40 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-vrg6z.spec.containers{with-cassandra-cassandra-schema-job} Started Started container with-cassandra-cassandra-schema-job kubelet logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:46 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job Completed Job completed job-controller logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:47 +0000 UTC Normal Pod with-cassandra-5f9cbb89f6-w28kz Binding Scheduled Successfully assigned kuttl-test-prepared-finch/with-cassandra-5f9cbb89f6-w28kz to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:47 +0000 UTC Normal Pod with-cassandra-5f9cbb89f6-w28kz AddedInterface Add eth0 [10.129.2.32/23] from ovn-kubernetes multus logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:47 +0000 UTC Normal Pod with-cassandra-5f9cbb89f6-w28kz.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" already present on machine kubelet logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:47 +0000 UTC Normal ReplicaSet.apps with-cassandra-5f9cbb89f6 SuccessfulCreate Created pod: with-cassandra-5f9cbb89f6-w28kz replicaset-controller logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:47 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-5f9cbb89f6 to 1 deployment-controller logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:48 +0000 UTC Normal Pod with-cassandra-5f9cbb89f6-w28kz.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:48 +0000 UTC Normal Pod with-cassandra-5f9cbb89f6-w28kz.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:48 +0000 UTC Normal Pod with-cassandra-5f9cbb89f6-w28kz.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:48 +0000 UTC Normal Pod with-cassandra-5f9cbb89f6-w28kz.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:48 +0000 UTC Normal Pod with-cassandra-5f9cbb89f6-w28kz.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:54 +0000 UTC Normal Pod with-cassandra-5f9cbb89f6-w28kz.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:54 +0000 UTC Normal Pod with-cassandra-5f9cbb89f6-w28kz.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:54 +0000 UTC Normal ReplicaSet.apps with-cassandra-5f9cbb89f6 SuccessfulDelete Deleted pod: with-cassandra-5f9cbb89f6-w28kz replicaset-controller logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:54 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled down replica set with-cassandra-5f9cbb89f6 to 0 from 1 deployment-controller logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:54 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-78bc6db7d9 to 1 deployment-controller logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:55 +0000 UTC Normal Pod with-cassandra-78bc6db7d9-k9jf2 Binding Scheduled Successfully assigned kuttl-test-prepared-finch/with-cassandra-78bc6db7d9-k9jf2 to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:55 +0000 UTC Normal Pod with-cassandra-78bc6db7d9-k9jf2 AddedInterface Add eth0 [10.129.2.33/23] from ovn-kubernetes multus logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:55 +0000 UTC Normal Pod with-cassandra-78bc6db7d9-k9jf2.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" already present on machine kubelet logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:55 +0000 UTC Normal Pod with-cassandra-78bc6db7d9-k9jf2.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:55 +0000 UTC Normal Pod with-cassandra-78bc6db7d9-k9jf2.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:55 +0000 UTC Normal Pod with-cassandra-78bc6db7d9-k9jf2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:55 +0000 UTC Normal Pod with-cassandra-78bc6db7d9-k9jf2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:55 +0000 UTC Normal Pod with-cassandra-78bc6db7d9-k9jf2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:55 +0000 UTC Normal ReplicaSet.apps with-cassandra-78bc6db7d9 SuccessfulCreate Created pod: with-cassandra-78bc6db7d9-k9jf2 replicaset-controller logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:11:56 +0000 UTC Warning Pod with-cassandra-78bc6db7d9-k9jf2.spec.containers{jaeger} BackOff Back-off restarting failed container jaeger in pod with-cassandra-78bc6db7d9-k9jf2_kuttl-test-prepared-finch(7939180c-2547-43b9-a1a9-995af7f62d54) kubelet logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:12:40 +0000 UTC Normal Pod check-span-zlmms Binding Scheduled Successfully assigned kuttl-test-prepared-finch/check-span-zlmms to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:12:40 +0000 UTC Normal Pod check-span-zlmms AddedInterface Add eth0 [10.131.0.71/23] from ovn-kubernetes multus logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:12:40 +0000 UTC Normal Pod check-span-zlmms.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:12:40 +0000 UTC Normal Pod check-span-zlmms.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:12:40 +0000 UTC Normal Pod check-span-zlmms.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:12:40 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-zlmms job-controller logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:12:40 +0000 UTC Normal Pod report-span-7k686 Binding Scheduled Successfully assigned kuttl-test-prepared-finch/report-span-7k686 to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:12:40 +0000 UTC Normal Pod report-span-7k686 AddedInterface Add eth0 [10.128.2.57/23] from ovn-kubernetes multus logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:12:40 +0000 UTC Normal Pod report-span-7k686.spec.containers{report-span} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:12:40 +0000 UTC Normal Pod report-span-7k686.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:12:40 +0000 UTC Normal Pod report-span-7k686.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:12:40 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-7k686 job-controller logger.go:42: 08:12:51 | examples-with-cassandra | 2024-05-20 08:12:51 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 08:12:51 | examples-with-cassandra | Deleting namespace: kuttl-test-prepared-finch === CONT kuttl/harness/examples-with-badger logger.go:42: 08:13:03 | examples-with-badger | Creating namespace: kuttl-test-knowing-antelope logger.go:42: 08:13:03 | examples-with-badger/0-install | starting test step 0-install logger.go:42: 08:13:03 | examples-with-badger/0-install | Jaeger:kuttl-test-knowing-antelope/with-badger created logger.go:42: 08:13:09 | examples-with-badger/0-install | test step completed 0-install logger.go:42: 08:13:09 | examples-with-badger/1-smoke-test | starting test step 1-smoke-test logger.go:42: 08:13:09 | examples-with-badger/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-badger /dev/null] logger.go:42: 08:13:11 | examples-with-badger/1-smoke-test | Warning: resource jaegers/with-badger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 08:13:17 | examples-with-badger/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337 JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 08:13:18 | examples-with-badger/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 08:13:18 | examples-with-badger/1-smoke-test | job.batch/report-span created logger.go:42: 08:13:18 | examples-with-badger/1-smoke-test | job.batch/check-span created logger.go:42: 08:13:30 | examples-with-badger/1-smoke-test | test step completed 1-smoke-test logger.go:42: 08:13:30 | examples-with-badger | examples-with-badger events from ns kuttl-test-knowing-antelope: logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:07 +0000 UTC Normal Pod with-badger-5687c558c-978f2 Binding Scheduled Successfully assigned kuttl-test-knowing-antelope/with-badger-5687c558c-978f2 to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:07 +0000 UTC Normal Pod with-badger-5687c558c-978f2 AddedInterface Add eth0 [10.129.2.34/23] from ovn-kubernetes multus logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:07 +0000 UTC Normal Pod with-badger-5687c558c-978f2.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" already present on machine kubelet logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:07 +0000 UTC Normal ReplicaSet.apps with-badger-5687c558c SuccessfulCreate Created pod: with-badger-5687c558c-978f2 replicaset-controller logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:07 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-5687c558c to 1 deployment-controller logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:08 +0000 UTC Normal Pod with-badger-5687c558c-978f2.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:08 +0000 UTC Normal Pod with-badger-5687c558c-978f2.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:08 +0000 UTC Normal Pod with-badger-5687c558c-978f2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:08 +0000 UTC Normal Pod with-badger-5687c558c-978f2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:08 +0000 UTC Normal Pod with-badger-5687c558c-978f2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:12 +0000 UTC Normal Pod with-badger-5687c558c-978f2.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:12 +0000 UTC Normal Pod with-badger-5687c558c-978f2.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:12 +0000 UTC Normal ReplicaSet.apps with-badger-5687c558c SuccessfulDelete Deleted pod: with-badger-5687c558c-978f2 replicaset-controller logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:12 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled down replica set with-badger-5687c558c to 0 from 1 deployment-controller logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:13 +0000 UTC Normal Pod with-badger-6cbb56ff4b-mnxb7 Binding Scheduled Successfully assigned kuttl-test-knowing-antelope/with-badger-6cbb56ff4b-mnxb7 to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:13 +0000 UTC Normal Pod with-badger-6cbb56ff4b-mnxb7 AddedInterface Add eth0 [10.129.2.35/23] from ovn-kubernetes multus logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:13 +0000 UTC Normal Pod with-badger-6cbb56ff4b-mnxb7.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" already present on machine kubelet logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:13 +0000 UTC Normal Pod with-badger-6cbb56ff4b-mnxb7.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:13 +0000 UTC Normal Pod with-badger-6cbb56ff4b-mnxb7.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:13 +0000 UTC Normal Pod with-badger-6cbb56ff4b-mnxb7.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:13 +0000 UTC Normal ReplicaSet.apps with-badger-6cbb56ff4b SuccessfulCreate Created pod: with-badger-6cbb56ff4b-mnxb7 replicaset-controller logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:13 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-6cbb56ff4b to 1 deployment-controller logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:14 +0000 UTC Normal Pod with-badger-6cbb56ff4b-mnxb7.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:14 +0000 UTC Normal Pod with-badger-6cbb56ff4b-mnxb7.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:18 +0000 UTC Normal Pod check-span-dx72w Binding Scheduled Successfully assigned kuttl-test-knowing-antelope/check-span-dx72w to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:18 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-dx72w job-controller logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:18 +0000 UTC Normal Pod report-span-xp6xt Binding Scheduled Successfully assigned kuttl-test-knowing-antelope/report-span-xp6xt to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:18 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-xp6xt job-controller logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:19 +0000 UTC Normal Pod check-span-dx72w AddedInterface Add eth0 [10.128.2.58/23] from ovn-kubernetes multus logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:19 +0000 UTC Normal Pod check-span-dx72w.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:19 +0000 UTC Normal Pod check-span-dx72w.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:19 +0000 UTC Normal Pod check-span-dx72w.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:19 +0000 UTC Normal Pod report-span-xp6xt AddedInterface Add eth0 [10.131.0.72/23] from ovn-kubernetes multus logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:19 +0000 UTC Normal Pod report-span-xp6xt.spec.containers{report-span} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:19 +0000 UTC Normal Pod report-span-xp6xt.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:19 +0000 UTC Normal Pod report-span-xp6xt.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 08:13:30 | examples-with-badger | 2024-05-20 08:13:29 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 08:13:30 | examples-with-badger | Deleting namespace: kuttl-test-knowing-antelope === CONT kuttl/harness/examples-simplest logger.go:42: 08:13:41 | examples-simplest | Creating namespace: kuttl-test-epic-ox logger.go:42: 08:13:41 | examples-simplest/0-install | starting test step 0-install logger.go:42: 08:13:42 | examples-simplest/0-install | Jaeger:kuttl-test-epic-ox/simplest created logger.go:42: 08:13:48 | examples-simplest/0-install | test step completed 0-install logger.go:42: 08:13:48 | examples-simplest/1-smoke-test | starting test step 1-smoke-test logger.go:42: 08:13:48 | examples-simplest/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 08:13:49 | examples-simplest/1-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 08:13:57 | examples-simplest/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337 JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 08:13:57 | examples-simplest/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 08:13:58 | examples-simplest/1-smoke-test | job.batch/report-span created logger.go:42: 08:13:58 | examples-simplest/1-smoke-test | job.batch/check-span created logger.go:42: 08:14:10 | examples-simplest/1-smoke-test | test step completed 1-smoke-test logger.go:42: 08:14:10 | examples-simplest | examples-simplest events from ns kuttl-test-epic-ox: logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:45 +0000 UTC Normal Pod simplest-68b7fd6cf5-qcn5r Binding Scheduled Successfully assigned kuttl-test-epic-ox/simplest-68b7fd6cf5-qcn5r to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:45 +0000 UTC Normal ReplicaSet.apps simplest-68b7fd6cf5 SuccessfulCreate Created pod: simplest-68b7fd6cf5-qcn5r replicaset-controller logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:45 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-68b7fd6cf5 to 1 deployment-controller logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:46 +0000 UTC Normal Pod simplest-68b7fd6cf5-qcn5r AddedInterface Add eth0 [10.129.2.36/23] from ovn-kubernetes multus logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:46 +0000 UTC Normal Pod simplest-68b7fd6cf5-qcn5r.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" already present on machine kubelet logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:46 +0000 UTC Normal Pod simplest-68b7fd6cf5-qcn5r.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:46 +0000 UTC Normal Pod simplest-68b7fd6cf5-qcn5r.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:46 +0000 UTC Normal Pod simplest-68b7fd6cf5-qcn5r.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:46 +0000 UTC Normal Pod simplest-68b7fd6cf5-qcn5r.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:46 +0000 UTC Normal Pod simplest-68b7fd6cf5-qcn5r.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:53 +0000 UTC Normal Pod simplest-68b7fd6cf5-qcn5r.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:53 +0000 UTC Normal Pod simplest-68b7fd6cf5-qcn5r.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:53 +0000 UTC Normal ReplicaSet.apps simplest-68b7fd6cf5 SuccessfulDelete Deleted pod: simplest-68b7fd6cf5-qcn5r replicaset-controller logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:53 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-68b7fd6cf5 to 0 from 1 deployment-controller logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:54 +0000 UTC Normal Pod simplest-67899895b5-nrttt Binding Scheduled Successfully assigned kuttl-test-epic-ox/simplest-67899895b5-nrttt to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:54 +0000 UTC Normal ReplicaSet.apps simplest-67899895b5 SuccessfulCreate Created pod: simplest-67899895b5-nrttt replicaset-controller logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:54 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-67899895b5 to 1 deployment-controller logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:55 +0000 UTC Normal Pod simplest-67899895b5-nrttt AddedInterface Add eth0 [10.129.2.37/23] from ovn-kubernetes multus logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:55 +0000 UTC Normal Pod simplest-67899895b5-nrttt.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" already present on machine kubelet logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:55 +0000 UTC Normal Pod simplest-67899895b5-nrttt.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:55 +0000 UTC Normal Pod simplest-67899895b5-nrttt.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:55 +0000 UTC Normal Pod simplest-67899895b5-nrttt.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:55 +0000 UTC Normal Pod simplest-67899895b5-nrttt.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:55 +0000 UTC Normal Pod simplest-67899895b5-nrttt.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:58 +0000 UTC Normal Pod check-span-zzqv8 Binding Scheduled Successfully assigned kuttl-test-epic-ox/check-span-zzqv8 to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:58 +0000 UTC Normal Pod check-span-zzqv8 AddedInterface Add eth0 [10.128.2.59/23] from ovn-kubernetes multus logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:58 +0000 UTC Normal Pod check-span-zzqv8.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:58 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-zzqv8 job-controller logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:58 +0000 UTC Normal Pod report-span-hkl29 Binding Scheduled Successfully assigned kuttl-test-epic-ox/report-span-hkl29 to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:58 +0000 UTC Normal Pod report-span-hkl29 AddedInterface Add eth0 [10.131.0.73/23] from ovn-kubernetes multus logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:58 +0000 UTC Normal Pod report-span-hkl29.spec.containers{report-span} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:58 +0000 UTC Normal Pod report-span-hkl29.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:58 +0000 UTC Normal Pod report-span-hkl29.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:58 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-hkl29 job-controller logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:59 +0000 UTC Normal Pod check-span-zzqv8.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:13:59 +0000 UTC Normal Pod check-span-zzqv8.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:14:10 | examples-simplest | 2024-05-20 08:14:09 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 08:14:10 | examples-simplest | Deleting namespace: kuttl-test-epic-ox === CONT kuttl/harness/examples-simple-prod-with-volumes logger.go:42: 08:14:22 | examples-simple-prod-with-volumes | Ignoring 03-check-volume.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:14:22 | examples-simple-prod-with-volumes | Creating namespace: kuttl-test-mighty-malamute logger.go:42: 08:14:22 | examples-simple-prod-with-volumes/1-install | starting test step 1-install logger.go:42: 08:14:22 | examples-simple-prod-with-volumes/1-install | Jaeger:kuttl-test-mighty-malamute/simple-prod created logger.go:42: 08:14:59 | examples-simple-prod-with-volumes/1-install | test step completed 1-install logger.go:42: 08:14:59 | examples-simple-prod-with-volumes/2-smoke-test | starting test step 2-smoke-test logger.go:42: 08:14:59 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 08:15:00 | examples-simple-prod-with-volumes/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 08:15:08 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337 JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 08:15:09 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 08:15:09 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/report-span created logger.go:42: 08:15:09 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/check-span created logger.go:42: 08:15:22 | examples-simple-prod-with-volumes/2-smoke-test | test step completed 2-smoke-test logger.go:42: 08:15:22 | examples-simple-prod-with-volumes/3-check-volume | starting test step 3-check-volume logger.go:42: 08:15:22 | examples-simple-prod-with-volumes/3-check-volume | running command: [sh -c kubectl exec $(kubectl get pods -n $NAMESPACE -l app=jaeger -l app.kubernetes.io/component=collector -o yaml | /tmp/jaeger-tests/bin/yq e '.items[0].metadata.name') -n $NAMESPACE -- ls /usr/share/elasticsearch/data] logger.go:42: 08:15:22 | examples-simple-prod-with-volumes/3-check-volume | test step completed 3-check-volume logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | examples-simple-prod-with-volumes events from ns kuttl-test-mighty-malamute: logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:29 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestmightymalamutesimpleprod-1-5b8cf8cff6 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestmightymalamutesimpleprod-1-5b8cks7m4 replicaset-controller logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmightymalamutesimpleprod-1-5b8cks7m4 Binding Scheduled Successfully assigned kuttl-test-mighty-malamute/elasticsearch-cdm-kuttltestmightymalamutesimpleprod-1-5b8cks7m4 to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmightymalamutesimpleprod-1-5b8cks7m4 AddedInterface Add eth0 [10.129.2.38/23] from ovn-kubernetes multus logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmightymalamutesimpleprod-1-5b8cks7m4.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmightymalamutesimpleprod-1-5b8cks7m4.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmightymalamutesimpleprod-1-5b8cks7m4.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmightymalamutesimpleprod-1-5b8cks7m4.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:29 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestmightymalamutesimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestmightymalamutesimpleprod-1-5b8cf8cff6 to 1 deployment-controller logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:30 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmightymalamutesimpleprod-1-5b8cks7m4.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:30 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmightymalamutesimpleprod-1-5b8cks7m4.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:39 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmightymalamutesimpleprod-1-5b8cks7m4.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:44 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmightymalamutesimpleprod-1-5b8cks7m4.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:56 +0000 UTC Normal Pod simple-prod-collector-574dd86499-tsrbk Binding Scheduled Successfully assigned kuttl-test-mighty-malamute/simple-prod-collector-574dd86499-tsrbk to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:56 +0000 UTC Normal Pod simple-prod-collector-574dd86499-tsrbk AddedInterface Add eth0 [10.131.0.74/23] from ovn-kubernetes multus logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:56 +0000 UTC Normal Pod simple-prod-collector-574dd86499-tsrbk.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:261a2c64c771323ad259e14802703bb790ee6b486b72c4939b487fbd83c30703" already present on machine kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:56 +0000 UTC Normal Pod simple-prod-collector-574dd86499-tsrbk.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:56 +0000 UTC Normal Pod simple-prod-collector-574dd86499-tsrbk.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:56 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-574dd86499 SuccessfulCreate Created pod: simple-prod-collector-574dd86499-tsrbk replicaset-controller logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:56 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-574dd86499 to 1 deployment-controller logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:56 +0000 UTC Normal Pod simple-prod-query-5c8598bcb4-7l7p5 Binding Scheduled Successfully assigned kuttl-test-mighty-malamute/simple-prod-query-5c8598bcb4-7l7p5 to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:56 +0000 UTC Normal Pod simple-prod-query-5c8598bcb4-7l7p5 AddedInterface Add eth0 [10.128.2.60/23] from ovn-kubernetes multus logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:56 +0000 UTC Normal Pod simple-prod-query-5c8598bcb4-7l7p5.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:156ea63cbd4dcd8c44173f118dba74388adbf1e82493237a4c514bb14131d5b8" already present on machine kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:56 +0000 UTC Normal Pod simple-prod-query-5c8598bcb4-7l7p5.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:56 +0000 UTC Normal Pod simple-prod-query-5c8598bcb4-7l7p5.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:56 +0000 UTC Normal ReplicaSet.apps simple-prod-query-5c8598bcb4 SuccessfulCreate Created pod: simple-prod-query-5c8598bcb4-7l7p5 replicaset-controller logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:56 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-5c8598bcb4 to 1 deployment-controller logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:57 +0000 UTC Normal Pod simple-prod-query-5c8598bcb4-7l7p5.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:57 +0000 UTC Normal Pod simple-prod-query-5c8598bcb4-7l7p5.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:57 +0000 UTC Normal Pod simple-prod-query-5c8598bcb4-7l7p5.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:57 +0000 UTC Normal Pod simple-prod-query-5c8598bcb4-7l7p5.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" already present on machine kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:57 +0000 UTC Normal Pod simple-prod-query-5c8598bcb4-7l7p5.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:14:57 +0000 UTC Normal Pod simple-prod-query-5c8598bcb4-7l7p5.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:04 +0000 UTC Normal Pod simple-prod-query-5c8598bcb4-7l7p5.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:04 +0000 UTC Normal Pod simple-prod-query-5c8598bcb4-7l7p5.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:04 +0000 UTC Normal Pod simple-prod-query-5c8598bcb4-7l7p5.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:04 +0000 UTC Normal ReplicaSet.apps simple-prod-query-5c8598bcb4 SuccessfulDelete Deleted pod: simple-prod-query-5c8598bcb4-7l7p5 replicaset-controller logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:04 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-5c8598bcb4 to 0 from 1 deployment-controller logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:05 +0000 UTC Normal Pod simple-prod-query-bcb9f687f-jvw67 Binding Scheduled Successfully assigned kuttl-test-mighty-malamute/simple-prod-query-bcb9f687f-jvw67 to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:05 +0000 UTC Normal ReplicaSet.apps simple-prod-query-bcb9f687f SuccessfulCreate Created pod: simple-prod-query-bcb9f687f-jvw67 replicaset-controller logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:05 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-bcb9f687f to 1 deployment-controller logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:06 +0000 UTC Normal Pod simple-prod-query-bcb9f687f-jvw67 AddedInterface Add eth0 [10.128.2.62/23] from ovn-kubernetes multus logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:06 +0000 UTC Normal Pod simple-prod-query-bcb9f687f-jvw67.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:156ea63cbd4dcd8c44173f118dba74388adbf1e82493237a4c514bb14131d5b8" already present on machine kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:06 +0000 UTC Normal Pod simple-prod-query-bcb9f687f-jvw67.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:06 +0000 UTC Normal Pod simple-prod-query-bcb9f687f-jvw67.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:06 +0000 UTC Normal Pod simple-prod-query-bcb9f687f-jvw67.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:06 +0000 UTC Normal Pod simple-prod-query-bcb9f687f-jvw67.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:06 +0000 UTC Normal Pod simple-prod-query-bcb9f687f-jvw67.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:06 +0000 UTC Normal Pod simple-prod-query-bcb9f687f-jvw67.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" already present on machine kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:06 +0000 UTC Normal Pod simple-prod-query-bcb9f687f-jvw67.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:06 +0000 UTC Normal Pod simple-prod-query-bcb9f687f-jvw67.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:09 +0000 UTC Normal Pod check-span-mvm79 Binding Scheduled Successfully assigned kuttl-test-mighty-malamute/check-span-mvm79 to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:09 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-mvm79 job-controller logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:09 +0000 UTC Normal Pod report-span-lbwww Binding Scheduled Successfully assigned kuttl-test-mighty-malamute/report-span-lbwww to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:09 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-lbwww job-controller logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:10 +0000 UTC Normal Pod check-span-mvm79 AddedInterface Add eth0 [10.131.0.76/23] from ovn-kubernetes multus logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:10 +0000 UTC Normal Pod check-span-mvm79.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:10 +0000 UTC Normal Pod check-span-mvm79.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:10 +0000 UTC Normal Pod check-span-mvm79.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:10 +0000 UTC Normal Pod report-span-lbwww AddedInterface Add eth0 [10.131.0.75/23] from ovn-kubernetes multus logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:10 +0000 UTC Normal Pod report-span-lbwww.spec.containers{report-span} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:10 +0000 UTC Normal Pod report-span-lbwww.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:10 +0000 UTC Normal Pod report-span-lbwww.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:11 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:11 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod simple-prod-collector-574dd86499-tsrbk horizontal-pod-autoscaler logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:11 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | 2024-05-20 08:15:21 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 08:15:22 | examples-simple-prod-with-volumes | Deleting namespace: kuttl-test-mighty-malamute === CONT kuttl/harness/examples-business-application-injected-sidecar logger.go:42: 08:15:35 | examples-business-application-injected-sidecar | Creating namespace: kuttl-test-together-moose logger.go:42: 08:15:35 | examples-business-application-injected-sidecar/0-install | starting test step 0-install logger.go:42: 08:15:35 | examples-business-application-injected-sidecar/0-install | Deployment:kuttl-test-together-moose/myapp created logger.go:42: 08:15:35 | examples-business-application-injected-sidecar/0-install | test step completed 0-install logger.go:42: 08:15:35 | examples-business-application-injected-sidecar/1-install | starting test step 1-install logger.go:42: 08:15:35 | examples-business-application-injected-sidecar/1-install | Jaeger:kuttl-test-together-moose/simplest created logger.go:42: 08:15:50 | examples-business-application-injected-sidecar/1-install | test step completed 1-install logger.go:42: 08:15:50 | examples-business-application-injected-sidecar/2-smoke-test | starting test step 2-smoke-test logger.go:42: 08:15:50 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 08:15:52 | examples-business-application-injected-sidecar/2-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 08:15:58 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337 JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 08:15:59 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 08:15:59 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/report-span created logger.go:42: 08:15:59 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/check-span created logger.go:42: 08:16:11 | examples-business-application-injected-sidecar/2-smoke-test | test step completed 2-smoke-test logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | examples-business-application-injected-sidecar events from ns kuttl-test-together-moose: logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:35 +0000 UTC Normal Pod myapp-679f79d5f8-gpn7f Binding Scheduled Successfully assigned kuttl-test-together-moose/myapp-679f79d5f8-gpn7f to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:35 +0000 UTC Normal ReplicaSet.apps myapp-679f79d5f8 SuccessfulCreate Created pod: myapp-679f79d5f8-gpn7f replicaset-controller logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:35 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-679f79d5f8 to 1 deployment-controller logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:36 +0000 UTC Warning Pod myapp-679f79d5f8-gpn7f FailedMount MountVolume.SetUp failed for volume "kube-api-access-mztvt" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:37 +0000 UTC Normal Pod myapp-679f79d5f8-gpn7f AddedInterface Add eth0 [10.131.0.77/23] from ovn-kubernetes multus logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:37 +0000 UTC Normal Pod myapp-679f79d5f8-gpn7f.spec.containers{myapp} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:38 +0000 UTC Normal Pod myapp-68db6d586c-b9jh5 Binding Scheduled Successfully assigned kuttl-test-together-moose/myapp-68db6d586c-b9jh5 to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:38 +0000 UTC Normal ReplicaSet.apps myapp-68db6d586c SuccessfulCreate Created pod: myapp-68db6d586c-b9jh5 replicaset-controller logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:38 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-68db6d586c to 1 deployment-controller logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:39 +0000 UTC Warning Pod myapp-68db6d586c-b9jh5 FailedMount MountVolume.SetUp failed for volume "simplest-trusted-ca" : configmap "simplest-trusted-ca" not found kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:39 +0000 UTC Warning Pod myapp-68db6d586c-b9jh5 FailedMount MountVolume.SetUp failed for volume "simplest-service-ca" : configmap "simplest-service-ca" not found kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:41 +0000 UTC Normal Pod myapp-679f79d5f8-gpn7f.spec.containers{myapp} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 3.792s (3.792s including waiting) kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:41 +0000 UTC Normal Pod myapp-679f79d5f8-gpn7f.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:41 +0000 UTC Normal Pod myapp-679f79d5f8-gpn7f.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:46 +0000 UTC Normal Pod myapp-68db6d586c-b9jh5 AddedInterface Add eth0 [10.129.2.39/23] from ovn-kubernetes multus logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:46 +0000 UTC Normal Pod simplest-5c88d4f8bf-z4n28 Binding Scheduled Successfully assigned kuttl-test-together-moose/simplest-5c88d4f8bf-z4n28 to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:46 +0000 UTC Normal Pod simplest-5c88d4f8bf-z4n28 AddedInterface Add eth0 [10.128.2.63/23] from ovn-kubernetes multus logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:46 +0000 UTC Normal ReplicaSet.apps simplest-5c88d4f8bf SuccessfulCreate Created pod: simplest-5c88d4f8bf-z4n28 replicaset-controller logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:46 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-5c88d4f8bf to 1 deployment-controller logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:47 +0000 UTC Normal Pod myapp-68db6d586c-b9jh5.spec.containers{myapp} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:47 +0000 UTC Normal Pod simplest-5c88d4f8bf-z4n28.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:48 +0000 UTC Warning Pod myapp-679f79d5f8-gpn7f.spec.containers{myapp} Unhealthy Liveness probe failed: Get "http://10.131.0.77:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:49 +0000 UTC Normal Pod simplest-5c88d4f8bf-z4n28.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" in 2.631s (2.631s including waiting) kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:49 +0000 UTC Normal Pod simplest-5c88d4f8bf-z4n28.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:49 +0000 UTC Normal Pod simplest-5c88d4f8bf-z4n28.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:49 +0000 UTC Normal Pod simplest-5c88d4f8bf-z4n28.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:49 +0000 UTC Normal Pod simplest-5c88d4f8bf-z4n28.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:49 +0000 UTC Normal Pod simplest-5c88d4f8bf-z4n28.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:51 +0000 UTC Normal Pod myapp-68db6d586c-b9jh5.spec.containers{myapp} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 4.204s (4.204s including waiting) kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:51 +0000 UTC Normal Pod myapp-68db6d586c-b9jh5.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:51 +0000 UTC Normal Pod myapp-68db6d586c-b9jh5.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:51 +0000 UTC Normal Pod myapp-68db6d586c-b9jh5.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:53 +0000 UTC Normal Pod myapp-68db6d586c-b9jh5.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" in 2.437s (2.437s including waiting) kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:53 +0000 UTC Normal Pod myapp-68db6d586c-b9jh5.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:53 +0000 UTC Normal Pod myapp-68db6d586c-b9jh5.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:54 +0000 UTC Normal Pod myapp-679f79d5f8-gpn7f.spec.containers{myapp} Killing Stopping container myapp kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:54 +0000 UTC Normal ReplicaSet.apps myapp-679f79d5f8 SuccessfulDelete Deleted pod: myapp-679f79d5f8-gpn7f replicaset-controller logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:54 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled down replica set myapp-679f79d5f8 to 0 from 1 deployment-controller logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:57 +0000 UTC Warning Pod myapp-68db6d586c-b9jh5.spec.containers{myapp} Unhealthy Liveness probe failed: Get "http://10.129.2.39:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:58 +0000 UTC Normal Pod simplest-5c88d4f8bf-z4n28.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:58 +0000 UTC Normal Pod simplest-5c88d4f8bf-z4n28.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:58 +0000 UTC Normal ReplicaSet.apps simplest-5c88d4f8bf SuccessfulDelete Deleted pod: simplest-5c88d4f8bf-z4n28 replicaset-controller logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:58 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-5c88d4f8bf to 0 from 1 deployment-controller logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:59 +0000 UTC Normal Pod check-span-ldd6k Binding Scheduled Successfully assigned kuttl-test-together-moose/check-span-ldd6k to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:59 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-ldd6k job-controller logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:59 +0000 UTC Normal Pod report-span-2j4dr Binding Scheduled Successfully assigned kuttl-test-together-moose/report-span-2j4dr to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:15:59 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-2j4dr job-controller logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:16:00 +0000 UTC Normal Pod check-span-ldd6k AddedInterface Add eth0 [10.128.2.64/23] from ovn-kubernetes multus logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:16:00 +0000 UTC Normal Pod check-span-ldd6k.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:16:00 +0000 UTC Normal Pod check-span-ldd6k.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:16:00 +0000 UTC Normal Pod check-span-ldd6k.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:16:00 +0000 UTC Normal Pod report-span-2j4dr AddedInterface Add eth0 [10.131.0.78/23] from ovn-kubernetes multus logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:16:00 +0000 UTC Normal Pod report-span-2j4dr.spec.containers{report-span} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:16:00 +0000 UTC Normal Pod report-span-2j4dr.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:16:00 +0000 UTC Normal Pod report-span-2j4dr.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:16:00 +0000 UTC Normal Pod simplest-787c76456b-bw6v2 Binding Scheduled Successfully assigned kuttl-test-together-moose/simplest-787c76456b-bw6v2 to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:16:00 +0000 UTC Normal Pod simplest-787c76456b-bw6v2 AddedInterface Add eth0 [10.131.0.79/23] from ovn-kubernetes multus logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:16:00 +0000 UTC Normal Pod simplest-787c76456b-bw6v2.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:16:00 +0000 UTC Normal ReplicaSet.apps simplest-787c76456b SuccessfulCreate Created pod: simplest-787c76456b-bw6v2 replicaset-controller logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:16:00 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-787c76456b to 1 deployment-controller logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:16:02 +0000 UTC Normal Pod simplest-787c76456b-bw6v2.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" in 1.675s (1.675s including waiting) kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:16:02 +0000 UTC Normal Pod simplest-787c76456b-bw6v2.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:16:02 +0000 UTC Normal Pod simplest-787c76456b-bw6v2.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:16:02 +0000 UTC Normal Pod simplest-787c76456b-bw6v2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:16:02 +0000 UTC Normal Pod simplest-787c76456b-bw6v2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:16:02 +0000 UTC Normal Pod simplest-787c76456b-bw6v2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | 2024-05-20 08:16:11 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 08:16:11 | examples-business-application-injected-sidecar | Deleting namespace: kuttl-test-together-moose === CONT kuttl/harness/examples-service-types logger.go:42: 08:16:24 | examples-service-types | Creating namespace: kuttl-test-enough-sheep logger.go:42: 08:16:24 | examples-service-types/0-install | starting test step 0-install logger.go:42: 08:16:24 | examples-service-types/0-install | Jaeger:kuttl-test-enough-sheep/service-types created logger.go:42: 08:16:30 | examples-service-types/0-install | test step completed 0-install logger.go:42: 08:16:30 | examples-service-types/1-smoke-test | starting test step 1-smoke-test logger.go:42: 08:16:30 | examples-service-types/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE service-types /dev/null] logger.go:42: 08:16:32 | examples-service-types/1-smoke-test | Warning: resource jaegers/service-types is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 08:16:38 | examples-service-types/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337 JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://service-types-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 08:16:39 | examples-service-types/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 08:16:39 | examples-service-types/1-smoke-test | job.batch/report-span created logger.go:42: 08:16:39 | examples-service-types/1-smoke-test | job.batch/check-span created logger.go:42: 08:16:51 | examples-service-types/1-smoke-test | test step completed 1-smoke-test logger.go:42: 08:16:51 | examples-service-types/2- | starting test step 2- logger.go:42: 08:16:52 | examples-service-types/2- | test step completed 2- logger.go:42: 08:16:52 | examples-service-types | examples-service-types events from ns kuttl-test-enough-sheep: logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:27 +0000 UTC Normal Service service-types-collector EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:28 +0000 UTC Normal Pod service-types-5856456fcf-h2lcp Binding Scheduled Successfully assigned kuttl-test-enough-sheep/service-types-5856456fcf-h2lcp to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:28 +0000 UTC Normal Pod service-types-5856456fcf-h2lcp AddedInterface Add eth0 [10.129.2.40/23] from ovn-kubernetes multus logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:28 +0000 UTC Normal Pod service-types-5856456fcf-h2lcp.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" already present on machine kubelet logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:28 +0000 UTC Normal Pod service-types-5856456fcf-h2lcp.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:28 +0000 UTC Normal Pod service-types-5856456fcf-h2lcp.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:28 +0000 UTC Normal Pod service-types-5856456fcf-h2lcp.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:28 +0000 UTC Normal Pod service-types-5856456fcf-h2lcp.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:28 +0000 UTC Normal Pod service-types-5856456fcf-h2lcp.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:28 +0000 UTC Normal ReplicaSet.apps service-types-5856456fcf SuccessfulCreate Created pod: service-types-5856456fcf-h2lcp replicaset-controller logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:28 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-5856456fcf to 1 deployment-controller logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:30 +0000 UTC Normal Service service-types-collector EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:30 +0000 UTC Normal Service service-types-query EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:33 +0000 UTC Normal Service service-types-query EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:36 +0000 UTC Normal Pod service-types-5856456fcf-h2lcp.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:36 +0000 UTC Normal Pod service-types-5856456fcf-h2lcp.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:36 +0000 UTC Normal ReplicaSet.apps service-types-5856456fcf SuccessfulDelete Deleted pod: service-types-5856456fcf-h2lcp replicaset-controller logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:36 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled down replica set service-types-5856456fcf to 0 from 1 deployment-controller logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:36 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-5674689866 to 1 deployment-controller logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:37 +0000 UTC Normal Pod service-types-5674689866-k5lld Binding Scheduled Successfully assigned kuttl-test-enough-sheep/service-types-5674689866-k5lld to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:37 +0000 UTC Normal Pod service-types-5674689866-k5lld AddedInterface Add eth0 [10.129.2.41/23] from ovn-kubernetes multus logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:37 +0000 UTC Normal Pod service-types-5674689866-k5lld.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" already present on machine kubelet logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:37 +0000 UTC Normal Pod service-types-5674689866-k5lld.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:37 +0000 UTC Normal Pod service-types-5674689866-k5lld.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:37 +0000 UTC Normal Pod service-types-5674689866-k5lld.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:37 +0000 UTC Normal Pod service-types-5674689866-k5lld.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:37 +0000 UTC Normal Pod service-types-5674689866-k5lld.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:37 +0000 UTC Normal ReplicaSet.apps service-types-5674689866 SuccessfulCreate Created pod: service-types-5674689866-k5lld replicaset-controller logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:39 +0000 UTC Normal Pod check-span-wnvt2 Binding Scheduled Successfully assigned kuttl-test-enough-sheep/check-span-wnvt2 to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:39 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-wnvt2 job-controller logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:39 +0000 UTC Normal Pod report-span-p9pp6 Binding Scheduled Successfully assigned kuttl-test-enough-sheep/report-span-p9pp6 to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:39 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-p9pp6 job-controller logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:40 +0000 UTC Normal Pod check-span-wnvt2 AddedInterface Add eth0 [10.128.2.65/23] from ovn-kubernetes multus logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:40 +0000 UTC Normal Pod check-span-wnvt2.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:40 +0000 UTC Normal Pod check-span-wnvt2.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:40 +0000 UTC Normal Pod check-span-wnvt2.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:40 +0000 UTC Normal Pod report-span-p9pp6 AddedInterface Add eth0 [10.131.0.80/23] from ovn-kubernetes multus logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:40 +0000 UTC Normal Pod report-span-p9pp6.spec.containers{report-span} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:40 +0000 UTC Normal Pod report-span-p9pp6.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:40 +0000 UTC Normal Pod report-span-p9pp6.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 08:16:52 | examples-service-types | 2024-05-20 08:16:51 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 08:16:52 | examples-service-types | Deleting namespace: kuttl-test-enough-sheep === CONT kuttl/harness/examples-openshift-with-htpasswd logger.go:42: 08:17:19 | examples-openshift-with-htpasswd | Ignoring 00-install.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:17:19 | examples-openshift-with-htpasswd | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:17:19 | examples-openshift-with-htpasswd | Creating namespace: kuttl-test-glad-deer logger.go:42: 08:17:19 | examples-openshift-with-htpasswd/0-install | starting test step 0-install logger.go:42: 08:17:19 | examples-openshift-with-htpasswd/0-install | Secret:kuttl-test-glad-deer/htpasswd created logger.go:42: 08:17:19 | examples-openshift-with-htpasswd/0-install | test step completed 0-install logger.go:42: 08:17:19 | examples-openshift-with-htpasswd/1-install | starting test step 1-install logger.go:42: 08:17:19 | examples-openshift-with-htpasswd/1-install | Jaeger:kuttl-test-glad-deer/with-htpasswd created logger.go:42: 08:17:25 | examples-openshift-with-htpasswd/1-install | test step completed 1-install logger.go:42: 08:17:25 | examples-openshift-with-htpasswd/2-check-unsecured | starting test step 2-check-unsecured logger.go:42: 08:17:25 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [./ensure-ingress-host.sh] logger.go:42: 08:17:25 | examples-openshift-with-htpasswd/2-check-unsecured | Checking the Ingress host value was populated logger.go:42: 08:17:25 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 0 logger.go:42: 08:17:25 | examples-openshift-with-htpasswd/2-check-unsecured | Hostname is with-htpasswd-kuttl-test-glad-deer.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com logger.go:42: 08:17:25 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 08:17:25 | examples-openshift-with-htpasswd/2-check-unsecured | Checking an expected HTTP response logger.go:42: 08:17:25 | examples-openshift-with-htpasswd/2-check-unsecured | Running in OpenShift logger.go:42: 08:17:25 | examples-openshift-with-htpasswd/2-check-unsecured | Not using any secret logger.go:42: 08:17:25 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 1/30 the https://with-htpasswd-kuttl-test-glad-deer.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 08:17:25 | examples-openshift-with-htpasswd/2-check-unsecured | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 08:17:25 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 2/30 the https://with-htpasswd-kuttl-test-glad-deer.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 08:17:26 | examples-openshift-with-htpasswd/2-check-unsecured | HTTP response is 503. 403 expected. Waiting 10 s logger.go:42: 08:17:36 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 3/30 the https://with-htpasswd-kuttl-test-glad-deer.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 08:17:36 | examples-openshift-with-htpasswd/2-check-unsecured | curl response asserted properly logger.go:42: 08:17:36 | examples-openshift-with-htpasswd/2-check-unsecured | test step completed 2-check-unsecured logger.go:42: 08:17:36 | examples-openshift-with-htpasswd/3-check-unauthorized | starting test step 3-check-unauthorized logger.go:42: 08:17:36 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [./ensure-ingress-host.sh] logger.go:42: 08:17:36 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking the Ingress host value was populated logger.go:42: 08:17:36 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 0 logger.go:42: 08:17:36 | examples-openshift-with-htpasswd/3-check-unauthorized | Hostname is with-htpasswd-kuttl-test-glad-deer.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com logger.go:42: 08:17:36 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [sh -c JAEGER_USERNAME=wronguser JAEGER_PASSWORD=wrongpassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 08:17:36 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking an expected HTTP response logger.go:42: 08:17:36 | examples-openshift-with-htpasswd/3-check-unauthorized | Running in OpenShift logger.go:42: 08:17:36 | examples-openshift-with-htpasswd/3-check-unauthorized | Using Jaeger basic authentication logger.go:42: 08:17:36 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 1/30 the https://with-htpasswd-kuttl-test-glad-deer.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 08:17:36 | examples-openshift-with-htpasswd/3-check-unauthorized | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 08:17:36 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 2/30 the https://with-htpasswd-kuttl-test-glad-deer.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 08:17:36 | examples-openshift-with-htpasswd/3-check-unauthorized | curl response asserted properly logger.go:42: 08:17:36 | examples-openshift-with-htpasswd/3-check-unauthorized | test step completed 3-check-unauthorized logger.go:42: 08:17:36 | examples-openshift-with-htpasswd/4-check-authorized | starting test step 4-check-authorized logger.go:42: 08:17:36 | examples-openshift-with-htpasswd/4-check-authorized | running command: [./ensure-ingress-host.sh] logger.go:42: 08:17:36 | examples-openshift-with-htpasswd/4-check-authorized | Checking the Ingress host value was populated logger.go:42: 08:17:36 | examples-openshift-with-htpasswd/4-check-authorized | Try number 0 logger.go:42: 08:17:36 | examples-openshift-with-htpasswd/4-check-authorized | Hostname is with-htpasswd-kuttl-test-glad-deer.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com logger.go:42: 08:17:36 | examples-openshift-with-htpasswd/4-check-authorized | running command: [sh -c JAEGER_USERNAME=awesomeuser JAEGER_PASSWORD=awesomepassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE with-htpasswd] logger.go:42: 08:17:36 | examples-openshift-with-htpasswd/4-check-authorized | Checking an expected HTTP response logger.go:42: 08:17:36 | examples-openshift-with-htpasswd/4-check-authorized | Running in OpenShift logger.go:42: 08:17:36 | examples-openshift-with-htpasswd/4-check-authorized | Using Jaeger basic authentication logger.go:42: 08:17:36 | examples-openshift-with-htpasswd/4-check-authorized | Try number 1/30 the https://with-htpasswd-kuttl-test-glad-deer.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 08:17:36 | examples-openshift-with-htpasswd/4-check-authorized | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 08:17:36 | examples-openshift-with-htpasswd/4-check-authorized | Try number 2/30 the https://with-htpasswd-kuttl-test-glad-deer.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 08:17:36 | examples-openshift-with-htpasswd/4-check-authorized | curl response asserted properly logger.go:42: 08:17:36 | examples-openshift-with-htpasswd/4-check-authorized | test step completed 4-check-authorized logger.go:42: 08:17:36 | examples-openshift-with-htpasswd | examples-openshift-with-htpasswd events from ns kuttl-test-glad-deer: logger.go:42: 08:17:36 | examples-openshift-with-htpasswd | 2024-05-20 08:17:23 +0000 UTC Normal Pod with-htpasswd-5dffd4db4c-dk792 Binding Scheduled Successfully assigned kuttl-test-glad-deer/with-htpasswd-5dffd4db4c-dk792 to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:17:36 | examples-openshift-with-htpasswd | 2024-05-20 08:17:23 +0000 UTC Normal Pod with-htpasswd-5dffd4db4c-dk792 AddedInterface Add eth0 [10.129.2.42/23] from ovn-kubernetes multus logger.go:42: 08:17:36 | examples-openshift-with-htpasswd | 2024-05-20 08:17:23 +0000 UTC Normal Pod with-htpasswd-5dffd4db4c-dk792.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" already present on machine kubelet logger.go:42: 08:17:36 | examples-openshift-with-htpasswd | 2024-05-20 08:17:23 +0000 UTC Normal Pod with-htpasswd-5dffd4db4c-dk792.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:17:36 | examples-openshift-with-htpasswd | 2024-05-20 08:17:23 +0000 UTC Normal Pod with-htpasswd-5dffd4db4c-dk792.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:17:36 | examples-openshift-with-htpasswd | 2024-05-20 08:17:23 +0000 UTC Normal Pod with-htpasswd-5dffd4db4c-dk792.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:17:36 | examples-openshift-with-htpasswd | 2024-05-20 08:17:23 +0000 UTC Normal Pod with-htpasswd-5dffd4db4c-dk792.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:17:36 | examples-openshift-with-htpasswd | 2024-05-20 08:17:23 +0000 UTC Normal Pod with-htpasswd-5dffd4db4c-dk792.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:17:36 | examples-openshift-with-htpasswd | 2024-05-20 08:17:23 +0000 UTC Normal ReplicaSet.apps with-htpasswd-5dffd4db4c SuccessfulCreate Created pod: with-htpasswd-5dffd4db4c-dk792 replicaset-controller logger.go:42: 08:17:36 | examples-openshift-with-htpasswd | 2024-05-20 08:17:23 +0000 UTC Normal Deployment.apps with-htpasswd ScalingReplicaSet Scaled up replica set with-htpasswd-5dffd4db4c to 1 deployment-controller logger.go:42: 08:17:36 | examples-openshift-with-htpasswd | Deleting namespace: kuttl-test-glad-deer === CONT kuttl/harness/examples-all-in-one-with-options logger.go:42: 08:17:42 | examples-all-in-one-with-options | Creating namespace: kuttl-test-leading-prawn logger.go:42: 08:17:42 | examples-all-in-one-with-options/0-install | starting test step 0-install logger.go:42: 08:17:43 | examples-all-in-one-with-options/0-install | Jaeger:kuttl-test-leading-prawn/my-jaeger created logger.go:42: 08:17:49 | examples-all-in-one-with-options/0-install | test step completed 0-install logger.go:42: 08:17:49 | examples-all-in-one-with-options/1-smoke-test | starting test step 1-smoke-test logger.go:42: 08:17:49 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 08:17:50 | examples-all-in-one-with-options/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 08:17:57 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c ASSERT_IMG=registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443/jaeger MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 08:17:57 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 08:17:58 | examples-all-in-one-with-options/1-smoke-test | job.batch/report-span created logger.go:42: 08:17:58 | examples-all-in-one-with-options/1-smoke-test | job.batch/check-span created logger.go:42: 08:18:10 | examples-all-in-one-with-options/1-smoke-test | test step completed 1-smoke-test logger.go:42: 08:18:10 | examples-all-in-one-with-options | examples-all-in-one-with-options events from ns kuttl-test-leading-prawn: logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:46 +0000 UTC Normal Pod my-jaeger-75968dfb4d-cx58d Binding Scheduled Successfully assigned kuttl-test-leading-prawn/my-jaeger-75968dfb4d-cx58d to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:46 +0000 UTC Normal ReplicaSet.apps my-jaeger-75968dfb4d SuccessfulCreate Created pod: my-jaeger-75968dfb4d-cx58d replicaset-controller logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:46 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-75968dfb4d to 1 deployment-controller logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:47 +0000 UTC Normal Pod my-jaeger-75968dfb4d-cx58d AddedInterface Add eth0 [10.129.2.43/23] from ovn-kubernetes multus logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:47 +0000 UTC Normal Pod my-jaeger-75968dfb4d-cx58d.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" already present on machine kubelet logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:47 +0000 UTC Normal Pod my-jaeger-75968dfb4d-cx58d.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:47 +0000 UTC Normal Pod my-jaeger-75968dfb4d-cx58d.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:47 +0000 UTC Normal Pod my-jaeger-75968dfb4d-cx58d.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:47 +0000 UTC Normal Pod my-jaeger-75968dfb4d-cx58d.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:47 +0000 UTC Normal Pod my-jaeger-75968dfb4d-cx58d.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:51 +0000 UTC Normal Pod my-jaeger-75968dfb4d-cx58d.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:51 +0000 UTC Normal Pod my-jaeger-75968dfb4d-cx58d.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:51 +0000 UTC Normal ReplicaSet.apps my-jaeger-75968dfb4d SuccessfulDelete Deleted pod: my-jaeger-75968dfb4d-cx58d replicaset-controller logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:51 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-75968dfb4d to 0 from 1 deployment-controller logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:52 +0000 UTC Normal Pod my-jaeger-87449f966-szfdd Binding Scheduled Successfully assigned kuttl-test-leading-prawn/my-jaeger-87449f966-szfdd to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:52 +0000 UTC Normal ReplicaSet.apps my-jaeger-87449f966 SuccessfulCreate Created pod: my-jaeger-87449f966-szfdd replicaset-controller logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:52 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-87449f966 to 1 deployment-controller logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:53 +0000 UTC Normal Pod my-jaeger-87449f966-szfdd AddedInterface Add eth0 [10.129.2.44/23] from ovn-kubernetes multus logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:53 +0000 UTC Normal Pod my-jaeger-87449f966-szfdd.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" already present on machine kubelet logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:53 +0000 UTC Normal Pod my-jaeger-87449f966-szfdd.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:53 +0000 UTC Normal Pod my-jaeger-87449f966-szfdd.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:53 +0000 UTC Normal Pod my-jaeger-87449f966-szfdd.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:53 +0000 UTC Normal Pod my-jaeger-87449f966-szfdd.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:53 +0000 UTC Normal Pod my-jaeger-87449f966-szfdd.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:58 +0000 UTC Normal Pod check-span-8h6l8 Binding Scheduled Successfully assigned kuttl-test-leading-prawn/check-span-8h6l8 to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:58 +0000 UTC Normal Pod check-span-8h6l8 AddedInterface Add eth0 [10.128.2.66/23] from ovn-kubernetes multus logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:58 +0000 UTC Normal Pod check-span-8h6l8.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:58 +0000 UTC Normal Pod check-span-8h6l8.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:58 +0000 UTC Normal Pod check-span-8h6l8.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:58 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-8h6l8 job-controller logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:58 +0000 UTC Normal Pod report-span-z9bgj Binding Scheduled Successfully assigned kuttl-test-leading-prawn/report-span-z9bgj to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:58 +0000 UTC Normal Pod report-span-z9bgj AddedInterface Add eth0 [10.131.0.81/23] from ovn-kubernetes multus logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:58 +0000 UTC Normal Pod report-span-z9bgj.spec.containers{report-span} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:58 +0000 UTC Normal Pod report-span-z9bgj.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:58 +0000 UTC Normal Pod report-span-z9bgj.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:17:58 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-z9bgj job-controller logger.go:42: 08:18:10 | examples-all-in-one-with-options | 2024-05-20 08:18:09 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 08:18:10 | examples-all-in-one-with-options | Deleting namespace: kuttl-test-leading-prawn === CONT kuttl/harness/examples-agent-with-priority-class logger.go:42: 08:18:22 | examples-agent-with-priority-class | Creating namespace: kuttl-test-welcome-manatee logger.go:42: 08:18:22 | examples-agent-with-priority-class/0-install | starting test step 0-install logger.go:42: 08:18:22 | examples-agent-with-priority-class/0-install | SecurityContextConstraints:/daemonset-with-hostport created logger.go:42: 08:18:22 | examples-agent-with-priority-class/0-install | ServiceAccount:kuttl-test-welcome-manatee/jaeger-agent-daemonset created logger.go:42: 08:18:22 | examples-agent-with-priority-class/0-install | test step completed 0-install logger.go:42: 08:18:22 | examples-agent-with-priority-class/1-add-policy | starting test step 1-add-policy logger.go:42: 08:18:22 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c oc adm policy --namespace $NAMESPACE add-scc-to-user daemonset-with-hostport -z jaeger-agent-daemonset] logger.go:42: 08:18:22 | examples-agent-with-priority-class/1-add-policy | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:daemonset-with-hostport added: "jaeger-agent-daemonset" logger.go:42: 08:18:22 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c sleep 5] logger.go:42: 08:18:27 | examples-agent-with-priority-class/1-add-policy | test step completed 1-add-policy logger.go:42: 08:18:27 | examples-agent-with-priority-class/2-install | starting test step 2-install logger.go:42: 08:18:27 | examples-agent-with-priority-class/2-install | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE agent-as-daemonset /dev/null] logger.go:42: 08:18:35 | examples-agent-with-priority-class/2-install | Error from server (NotFound): deployments.apps "agent-as-daemonset" not found logger.go:42: 08:18:35 | examples-agent-with-priority-class/2-install | command failure, skipping 2 additional commands logger.go:42: 08:18:35 | examples-agent-with-priority-class/2-install | PriorityClass:/high-priority created logger.go:42: 08:18:35 | examples-agent-with-priority-class/2-install | Jaeger:kuttl-test-welcome-manatee/agent-as-daemonset updated case.go:364: failed in step 2-install case.go:366: exit status 1 logger.go:42: 08:18:35 | examples-agent-with-priority-class | examples-agent-with-priority-class events from ns kuttl-test-welcome-manatee: logger.go:42: 08:18:35 | examples-agent-with-priority-class | Deleting namespace: kuttl-test-welcome-manatee === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- FAIL: kuttl (569.98s) --- FAIL: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.89s) --- PASS: kuttl/harness/examples-simple-prod (72.14s) --- PASS: kuttl/harness/examples-with-sampling (57.82s) --- PASS: kuttl/harness/examples-with-cassandra (95.79s) --- PASS: kuttl/harness/examples-with-badger (38.42s) --- PASS: kuttl/harness/examples-simplest (40.87s) --- PASS: kuttl/harness/examples-simple-prod-with-volumes (72.39s) --- PASS: kuttl/harness/examples-business-application-injected-sidecar (49.06s) --- PASS: kuttl/harness/examples-service-types (55.03s) --- PASS: kuttl/harness/examples-openshift-with-htpasswd (23.63s) --- PASS: kuttl/harness/examples-all-in-one-with-options (39.41s) --- FAIL: kuttl/harness/examples-agent-with-priority-class (19.37s) FAIL + exit_code=1 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name examples --report --output /logs/artifacts/examples.xml ./artifacts/kuttl-report.xml time="2024-05-20T08:18:41Z" level=debug msg="Setting a new name for the test suites" time="2024-05-20T08:18:41Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-05-20T08:18:41Z" level=debug msg="normalizing test case names" time="2024-05-20T08:18:41Z" level=debug msg="examples/artifacts -> examples_artifacts" time="2024-05-20T08:18:41Z" level=debug msg="examples/examples-simple-prod -> examples_examples_simple_prod" time="2024-05-20T08:18:41Z" level=debug msg="examples/examples-with-sampling -> examples_examples_with_sampling" time="2024-05-20T08:18:41Z" level=debug msg="examples/examples-with-cassandra -> examples_examples_with_cassandra" time="2024-05-20T08:18:41Z" level=debug msg="examples/examples-with-badger -> examples_examples_with_badger" time="2024-05-20T08:18:41Z" level=debug msg="examples/examples-simplest -> examples_examples_simplest" time="2024-05-20T08:18:41Z" level=debug msg="examples/examples-simple-prod-with-volumes -> examples_examples_simple_prod_with_volumes" time="2024-05-20T08:18:41Z" level=debug msg="examples/examples-business-application-injected-sidecar -> examples_examples_business_application_injected_sidecar" time="2024-05-20T08:18:41Z" level=debug msg="examples/examples-service-types -> examples_examples_service_types" time="2024-05-20T08:18:41Z" level=debug msg="examples/examples-openshift-with-htpasswd -> examples_examples_openshift_with_htpasswd" time="2024-05-20T08:18:41Z" level=debug msg="examples/examples-all-in-one-with-options -> examples_examples_all_in_one_with_options" time="2024-05-20T08:18:41Z" level=debug msg="examples/examples-agent-with-priority-class -> examples_examples_agent_with_priority_class" +---------------------------------------------------------+--------+ | NAME | RESULT | +---------------------------------------------------------+--------+ | examples_artifacts | passed | | examples_examples_simple_prod | passed | | examples_examples_with_sampling | passed | | examples_examples_with_cassandra | passed | | examples_examples_with_badger | passed | | examples_examples_simplest | passed | | examples_examples_simple_prod_with_volumes | passed | | examples_examples_business_application_injected_sidecar | passed | | examples_examples_service_types | passed | | examples_examples_openshift_with_htpasswd | passed | | examples_examples_all_in_one_with_options | passed | | examples_examples_agent_with_priority_class | failed | +---------------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh generate false true + '[' 3 -ne 3 ']' + test_suite_name=generate + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/generate.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-generate make[2]: Entering directory '/tmp/jaeger-tests' test -s /tmp/jaeger-tests/bin/operator-sdk || curl -sLo /tmp/jaeger-tests/bin/operator-sdk https://github.com/operator-framework/operator-sdk/releases/download/v1.32.0/operator-sdk_`go env GOOS`_`go env GOARCH` ./hack/install/install-golangci-lint.sh Installing golangci-lint golangci-lint 1.55.2 is installed already ./hack/install/install-goimports.sh Installing goimports Try 0... go install golang.org/x/tools/cmd/goimports@v0.1.12 >>>> Formatting code... ./.ci/format.sh >>>> Building... ./hack/install/install-dependencies.sh Installing go dependencies Try 0... go mod download GOOS= GOARCH= CGO_ENABLED=0 GO111MODULE=on go build -ldflags "-X "github.com/jaegertracing/jaeger-operator/pkg/version".version="1.57.0" -X "github.com/jaegertracing/jaeger-operator/pkg/version".buildDate=2024-05-20T08:18:42Z -X "github.com/jaegertracing/jaeger-operator/pkg/version".defaultJaeger="1.57.0"" -o "bin/jaeger-operator" main.go JAEGER_VERSION="1.57.0" ./tests/e2e/generate/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-05-19-083311 True False 37m Cluster version is 4.16.0-0.nightly-2024-05-19-083311' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-05-19-083311 True False 37m Cluster version is 4.16.0-0.nightly-2024-05-19-083311' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/generate/render.sh ++ export SUITE_DIR=./tests/e2e/generate ++ SUITE_DIR=./tests/e2e/generate ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/generate ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + '[' true = true ']' + skip_test generate 'This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 2 -ne 2 ']' + test_name=generate + message='This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/generate/_build + '[' _build '!=' _build ']' + rm -rf generate + warning 'generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed\e[0m' WAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running generate E2E tests' Running generate E2E tests + cd tests/e2e/generate/_build + set +e + KUBECONFIG=/tmp/kubeconfig-559878434 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 08:18:57 | artifacts | Creating namespace: kuttl-test-topical-perch logger.go:42: 08:18:58 | artifacts | artifacts events from ns kuttl-test-topical-perch: logger.go:42: 08:18:58 | artifacts | Deleting namespace: kuttl-test-topical-perch === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (6.50s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.34s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name generate --report --output /logs/artifacts/generate.xml ./artifacts/kuttl-report.xml time="2024-05-20T08:19:04Z" level=debug msg="Setting a new name for the test suites" time="2024-05-20T08:19:04Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-05-20T08:19:04Z" level=debug msg="normalizing test case names" time="2024-05-20T08:19:04Z" level=debug msg="generate/artifacts -> generate_artifacts" +--------------------+--------+ | NAME | RESULT | +--------------------+--------+ | generate_artifacts | passed | +--------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh upgrade false true + '[' 3 -ne 3 ']' + test_suite_name=upgrade + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/upgrade.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-upgrade make[2]: Entering directory '/tmp/jaeger-tests' make docker JAEGER_VERSION=1.57.1 IMG="quay.io//jaeger-operator:next" make[3]: Entering directory '/tmp/jaeger-tests' [ ! -z "true" ] || docker build --build-arg=GOPROXY= --build-arg=VERSION="1.57.0" --build-arg=JAEGER_VERSION=1.57.1 --build-arg=TARGETARCH= --build-arg VERSION_DATE=2024-05-20T08:19:04Z --build-arg VERSION_PKG="github.com/jaegertracing/jaeger-operator/pkg/version" -t "quay.io//jaeger-operator:next" . make[3]: Leaving directory '/tmp/jaeger-tests' touch build-e2e-upgrade-image SKIP_ES_EXTERNAL=true IMG=quay.io//jaeger-operator:"1.57.0" JAEGER_OPERATOR_VERSION="1.57.0" JAEGER_VERSION="1.57.0" ./tests/e2e/upgrade/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-05-19-083311 True False 37m Cluster version is 4.16.0-0.nightly-2024-05-19-083311' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-05-19-083311 True False 37m Cluster version is 4.16.0-0.nightly-2024-05-19-083311' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/upgrade/render.sh ++ export SUITE_DIR=./tests/e2e/upgrade ++ SUITE_DIR=./tests/e2e/upgrade ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/upgrade ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + export JAEGER_NAME + '[' true = true ']' + skip_test upgrade 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade + warning 'upgrade: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade: Test not supported in OpenShift\e[0m' WAR: upgrade: Test not supported in OpenShift + '[' true = true ']' + skip_test upgrade-from-latest-release 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade-from-latest-release + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade-from-latest-release + warning 'upgrade-from-latest-release: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade-from-latest-release: Test not supported in OpenShift\e[0m' WAR: upgrade-from-latest-release: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running upgrade E2E tests' Running upgrade E2E tests + cd tests/e2e/upgrade/_build + set +e + KUBECONFIG=/tmp/kubeconfig-559878434 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 08:19:05 | artifacts | Creating namespace: kuttl-test-poetic-boxer logger.go:42: 08:19:05 | artifacts | artifacts events from ns kuttl-test-poetic-boxer: logger.go:42: 08:19:05 | artifacts | Deleting namespace: kuttl-test-poetic-boxer === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (6.07s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.90s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name upgrade --report --output /logs/artifacts/upgrade.xml ./artifacts/kuttl-report.xml time="2024-05-20T08:19:11Z" level=debug msg="Setting a new name for the test suites" time="2024-05-20T08:19:11Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-05-20T08:19:11Z" level=debug msg="normalizing test case names" time="2024-05-20T08:19:11Z" level=debug msg="upgrade/artifacts -> upgrade_artifacts" +-------------------+--------+ | NAME | RESULT | +-------------------+--------+ | upgrade_artifacts | passed | +-------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh sidecar false true + '[' 3 -ne 3 ']' + test_suite_name=sidecar + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/sidecar.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-sidecar make[2]: Entering directory '/tmp/jaeger-tests' ./tests/e2e/sidecar/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-05-19-083311 True False 37m Cluster version is 4.16.0-0.nightly-2024-05-19-083311' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-05-19-083311 True False 37m Cluster version is 4.16.0-0.nightly-2024-05-19-083311' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/sidecar/render.sh ++ export SUITE_DIR=./tests/e2e/sidecar ++ SUITE_DIR=./tests/e2e/sidecar ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/sidecar ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + jaeger_service_name=order + start_test sidecar-deployment + '[' 1 -ne 1 ']' + test_name=sidecar-deployment + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-deployment' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-deployment\e[0m' Rendering files for test sidecar-deployment + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build + '[' _build '!=' _build ']' + mkdir -p sidecar-deployment + cd sidecar-deployment + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-namespace + '[' 1 -ne 1 ']' + test_name=sidecar-namespace + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-namespace' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-namespace\e[0m' Rendering files for test sidecar-namespace + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-deployment + '[' sidecar-deployment '!=' _build ']' + cd .. + mkdir -p sidecar-namespace + cd sidecar-namespace + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-skip-webhook + '[' 1 -ne 1 ']' + test_name=sidecar-skip-webhook + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-skip-webhook' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-skip-webhook\e[0m' Rendering files for test sidecar-skip-webhook + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-namespace + '[' sidecar-namespace '!=' _build ']' + cd .. + mkdir -p sidecar-skip-webhook + cd sidecar-skip-webhook + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running sidecar E2E tests' Running sidecar E2E tests + cd tests/e2e/sidecar/_build + set +e + KUBECONFIG=/tmp/kubeconfig-559878434 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 4 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/sidecar-deployment === PAUSE kuttl/harness/sidecar-deployment === RUN kuttl/harness/sidecar-namespace === PAUSE kuttl/harness/sidecar-namespace === RUN kuttl/harness/sidecar-skip-webhook === PAUSE kuttl/harness/sidecar-skip-webhook === CONT kuttl/harness/artifacts logger.go:42: 08:19:19 | artifacts | Creating namespace: kuttl-test-popular-minnow logger.go:42: 08:19:19 | artifacts | artifacts events from ns kuttl-test-popular-minnow: logger.go:42: 08:19:19 | artifacts | Deleting namespace: kuttl-test-popular-minnow === CONT kuttl/harness/sidecar-namespace logger.go:42: 08:19:25 | sidecar-namespace | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:19:25 | sidecar-namespace | Creating namespace: kuttl-test-immune-pangolin logger.go:42: 08:19:25 | sidecar-namespace/0-install | starting test step 0-install logger.go:42: 08:19:25 | sidecar-namespace/0-install | Jaeger:kuttl-test-immune-pangolin/agent-as-sidecar created logger.go:42: 08:19:32 | sidecar-namespace/0-install | test step completed 0-install logger.go:42: 08:19:32 | sidecar-namespace/1-install | starting test step 1-install logger.go:42: 08:19:32 | sidecar-namespace/1-install | Deployment:kuttl-test-immune-pangolin/vertx-create-span-sidecar created logger.go:42: 08:19:33 | sidecar-namespace/1-install | test step completed 1-install logger.go:42: 08:19:33 | sidecar-namespace/2-enable-injection | starting test step 2-enable-injection logger.go:42: 08:19:33 | sidecar-namespace/2-enable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="true"] logger.go:42: 08:19:33 | sidecar-namespace/2-enable-injection | namespace/kuttl-test-immune-pangolin annotated logger.go:42: 08:19:39 | sidecar-namespace/2-enable-injection | test step completed 2-enable-injection logger.go:42: 08:19:39 | sidecar-namespace/3-find-service | starting test step 3-find-service logger.go:42: 08:19:39 | sidecar-namespace/3-find-service | Job:kuttl-test-immune-pangolin/00-find-service created logger.go:42: 08:19:55 | sidecar-namespace/3-find-service | test step completed 3-find-service logger.go:42: 08:19:55 | sidecar-namespace/4-other-instance | starting test step 4-other-instance logger.go:42: 08:19:55 | sidecar-namespace/4-other-instance | Jaeger:kuttl-test-immune-pangolin/agent-as-sidecar2 created logger.go:42: 08:20:03 | sidecar-namespace/4-other-instance | test step completed 4-other-instance logger.go:42: 08:20:03 | sidecar-namespace/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 08:20:03 | sidecar-namespace/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 08:20:03 | sidecar-namespace/6-find-service | starting test step 6-find-service logger.go:42: 08:20:03 | sidecar-namespace/6-find-service | Job:kuttl-test-immune-pangolin/01-find-service created logger.go:42: 08:20:24 | sidecar-namespace/6-find-service | test step completed 6-find-service logger.go:42: 08:20:24 | sidecar-namespace/7-disable-injection | starting test step 7-disable-injection logger.go:42: 08:20:24 | sidecar-namespace/7-disable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="false"] logger.go:42: 08:20:24 | sidecar-namespace/7-disable-injection | namespace/kuttl-test-immune-pangolin annotated logger.go:42: 08:20:25 | sidecar-namespace/7-disable-injection | test step completed 7-disable-injection logger.go:42: 08:20:25 | sidecar-namespace | sidecar-namespace events from ns kuttl-test-immune-pangolin: logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:29 +0000 UTC Normal Pod agent-as-sidecar-7f7d4c76bd-hbpfb Binding Scheduled Successfully assigned kuttl-test-immune-pangolin/agent-as-sidecar-7f7d4c76bd-hbpfb to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:29 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-7f7d4c76bd SuccessfulCreate Created pod: agent-as-sidecar-7f7d4c76bd-hbpfb replicaset-controller logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:29 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-7f7d4c76bd to 1 deployment-controller logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:30 +0000 UTC Normal Pod agent-as-sidecar-7f7d4c76bd-hbpfb AddedInterface Add eth0 [10.129.2.45/23] from ovn-kubernetes multus logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:30 +0000 UTC Normal Pod agent-as-sidecar-7f7d4c76bd-hbpfb.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" already present on machine kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:30 +0000 UTC Normal Pod agent-as-sidecar-7f7d4c76bd-hbpfb.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:30 +0000 UTC Normal Pod agent-as-sidecar-7f7d4c76bd-hbpfb.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:32 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-gxqz8 Binding Scheduled Successfully assigned kuttl-test-immune-pangolin/vertx-create-span-sidecar-84d458b68c-gxqz8 to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:32 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-gxqz8 AddedInterface Add eth0 [10.131.0.82/23] from ovn-kubernetes multus logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:32 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-gxqz8.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:32 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-gxqz8.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:32 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-gxqz8.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:32 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-gxqz8 replicaset-controller logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:32 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:33 +0000 UTC Normal Pod vertx-create-span-sidecar-88d6fd77c-q7pml Binding Scheduled Successfully assigned kuttl-test-immune-pangolin/vertx-create-span-sidecar-88d6fd77c-q7pml to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:33 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-88d6fd77c SuccessfulCreate Created pod: vertx-create-span-sidecar-88d6fd77c-q7pml replicaset-controller logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:33 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-88d6fd77c to 1 deployment-controller logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:34 +0000 UTC Normal Pod vertx-create-span-sidecar-88d6fd77c-q7pml AddedInterface Add eth0 [10.128.2.67/23] from ovn-kubernetes multus logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:34 +0000 UTC Normal Pod vertx-create-span-sidecar-88d6fd77c-q7pml.spec.containers{vertx-create-span-sidecar} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:37 +0000 UTC Normal Pod vertx-create-span-sidecar-88d6fd77c-q7pml.spec.containers{vertx-create-span-sidecar} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 3.645s (3.645s including waiting) kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:37 +0000 UTC Normal Pod vertx-create-span-sidecar-88d6fd77c-q7pml.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:37 +0000 UTC Normal Pod vertx-create-span-sidecar-88d6fd77c-q7pml.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:37 +0000 UTC Normal Pod vertx-create-span-sidecar-88d6fd77c-q7pml.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" already present on machine kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:38 +0000 UTC Normal Pod vertx-create-span-sidecar-88d6fd77c-q7pml.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:38 +0000 UTC Normal Pod vertx-create-span-sidecar-88d6fd77c-q7pml.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:39 +0000 UTC Normal Pod 00-find-service-9sh6c Binding Scheduled Successfully assigned kuttl-test-immune-pangolin/00-find-service-9sh6c to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:39 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-9sh6c job-controller logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:40 +0000 UTC Normal Pod 00-find-service-9sh6c AddedInterface Add eth0 [10.129.2.46/23] from ovn-kubernetes multus logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:40 +0000 UTC Normal Pod 00-find-service-9sh6c.spec.containers{asserts-container} Pulling Pulling image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:40 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-gxqz8.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.82:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:40 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-gxqz8.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.131.0.82:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:42 +0000 UTC Normal Pod 00-find-service-9sh6c.spec.containers{asserts-container} Pulled Successfully pulled image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" in 2.543s (2.543s including waiting) kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:42 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-gxqz8.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:43 +0000 UTC Normal Pod 00-find-service-9sh6c.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:43 +0000 UTC Normal Pod 00-find-service-9sh6c.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:43 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-gxqz8.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.82:8080/": read tcp 10.131.0.2:55610->10.131.0.82:8080: read: connection reset by peer kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:43 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-gxqz8.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.82:8080/": dial tcp 10.131.0.82:8080: connect: connection refused kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:45 +0000 UTC Warning Pod vertx-create-span-sidecar-88d6fd77c-q7pml.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.128.2.67:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:45 +0000 UTC Warning Pod vertx-create-span-sidecar-88d6fd77c-q7pml.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.67:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:48 +0000 UTC Normal Pod vertx-create-span-sidecar-88d6fd77c-q7pml.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:48 +0000 UTC Warning Pod vertx-create-span-sidecar-88d6fd77c-q7pml.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.67:8080/": read tcp 10.128.2.2:46836->10.128.2.67:8080: read: connection reset by peer kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:48 +0000 UTC Warning Pod vertx-create-span-sidecar-88d6fd77c-q7pml.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.67:8080/": dial tcp 10.128.2.67:8080: connect: connection refused kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:48 +0000 UTC Normal Pod vertx-create-span-sidecar-88d6fd77c-q7pml.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:54 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:19:54 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-gxqz8.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.82:8080/": read tcp 10.131.0.2:39320->10.131.0.82:8080: read: connection reset by peer kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:01 +0000 UTC Normal Pod agent-as-sidecar2-6d5c577fc8-88ktc Binding Scheduled Successfully assigned kuttl-test-immune-pangolin/agent-as-sidecar2-6d5c577fc8-88ktc to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:01 +0000 UTC Normal Pod agent-as-sidecar2-6d5c577fc8-88ktc AddedInterface Add eth0 [10.129.2.47/23] from ovn-kubernetes multus logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:01 +0000 UTC Normal Pod agent-as-sidecar2-6d5c577fc8-88ktc.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" already present on machine kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:01 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-6d5c577fc8 SuccessfulCreate Created pod: agent-as-sidecar2-6d5c577fc8-88ktc replicaset-controller logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:01 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-6d5c577fc8 to 1 deployment-controller logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:02 +0000 UTC Normal Pod agent-as-sidecar2-6d5c577fc8-88ktc.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:02 +0000 UTC Normal Pod agent-as-sidecar2-6d5c577fc8-88ktc.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:03 +0000 UTC Normal Pod 01-find-service-rh5jm Binding Scheduled Successfully assigned kuttl-test-immune-pangolin/01-find-service-rh5jm to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:03 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-rh5jm job-controller logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:03 +0000 UTC Normal Pod agent-as-sidecar-7f7d4c76bd-hbpfb.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:04 +0000 UTC Normal Pod 01-find-service-rh5jm AddedInterface Add eth0 [10.131.0.83/23] from ovn-kubernetes multus logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:04 +0000 UTC Normal Pod 01-find-service-rh5jm.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:04 +0000 UTC Normal Pod 01-find-service-rh5jm.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:04 +0000 UTC Normal Pod 01-find-service-rh5jm.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:07 +0000 UTC Normal Pod vertx-create-span-sidecar-685b5d84bc-smjvs Binding Scheduled Successfully assigned kuttl-test-immune-pangolin/vertx-create-span-sidecar-685b5d84bc-smjvs to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:07 +0000 UTC Normal Pod vertx-create-span-sidecar-685b5d84bc-smjvs AddedInterface Add eth0 [10.129.2.48/23] from ovn-kubernetes multus logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:07 +0000 UTC Normal Pod vertx-create-span-sidecar-685b5d84bc-smjvs.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:07 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-685b5d84bc SuccessfulCreate Created pod: vertx-create-span-sidecar-685b5d84bc-smjvs replicaset-controller logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:07 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulDelete Deleted pod: vertx-create-span-sidecar-84d458b68c-gxqz8 replicaset-controller logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:07 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-84d458b68c to 0 from 1 deployment-controller logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:07 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-685b5d84bc to 1 from 0 deployment-controller logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:08 +0000 UTC Normal Pod vertx-create-span-sidecar-685b5d84bc-smjvs.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:08 +0000 UTC Normal Pod vertx-create-span-sidecar-685b5d84bc-smjvs.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:08 +0000 UTC Normal Pod vertx-create-span-sidecar-685b5d84bc-smjvs.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" already present on machine kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:08 +0000 UTC Normal Pod vertx-create-span-sidecar-685b5d84bc-smjvs.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:08 +0000 UTC Normal Pod vertx-create-span-sidecar-685b5d84bc-smjvs.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:16 +0000 UTC Warning Pod vertx-create-span-sidecar-685b5d84bc-smjvs.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.48:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:16 +0000 UTC Warning Pod vertx-create-span-sidecar-685b5d84bc-smjvs.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.48:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:18 +0000 UTC Normal Pod vertx-create-span-sidecar-685b5d84bc-smjvs.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:19 +0000 UTC Warning Pod vertx-create-span-sidecar-685b5d84bc-smjvs.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.48:8080/": read tcp 10.129.2.2:39534->10.129.2.48:8080: read: connection reset by peer kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:19 +0000 UTC Warning Pod vertx-create-span-sidecar-685b5d84bc-smjvs.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.48:8080/": dial tcp 10.129.2.48:8080: connect: connection refused kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:23 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:24 +0000 UTC Normal Pod vertx-create-span-sidecar-799b6796bd-4bzbl Binding Scheduled Successfully assigned kuttl-test-immune-pangolin/vertx-create-span-sidecar-799b6796bd-4bzbl to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:24 +0000 UTC Normal Pod vertx-create-span-sidecar-799b6796bd-4bzbl AddedInterface Add eth0 [10.131.0.84/23] from ovn-kubernetes multus logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:24 +0000 UTC Normal Pod vertx-create-span-sidecar-799b6796bd-4bzbl.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:24 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-799b6796bd SuccessfulCreate Created pod: vertx-create-span-sidecar-799b6796bd-4bzbl replicaset-controller logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:24 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-88d6fd77c SuccessfulDelete Deleted pod: vertx-create-span-sidecar-88d6fd77c-q7pml replicaset-controller logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:24 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-88d6fd77c to 0 from 1 deployment-controller logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:24 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-799b6796bd to 1 from 0 deployment-controller logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:25 +0000 UTC Normal Pod vertx-create-span-sidecar-799b6796bd-4bzbl.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 08:20:25 | sidecar-namespace | 2024-05-20 08:20:25 +0000 UTC Normal Pod vertx-create-span-sidecar-799b6796bd-4bzbl.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 08:20:25 | sidecar-namespace | Deleting namespace: kuttl-test-immune-pangolin === CONT kuttl/harness/sidecar-skip-webhook logger.go:42: 08:20:32 | sidecar-skip-webhook | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:20:32 | sidecar-skip-webhook | Creating namespace: kuttl-test-willing-barnacle logger.go:42: 08:20:32 | sidecar-skip-webhook/0-install | starting test step 0-install logger.go:42: 08:20:32 | sidecar-skip-webhook/0-install | Jaeger:kuttl-test-willing-barnacle/agent-as-sidecar created logger.go:42: 08:20:38 | sidecar-skip-webhook/0-install | test step completed 0-install logger.go:42: 08:20:38 | sidecar-skip-webhook/1-install | starting test step 1-install logger.go:42: 08:20:39 | sidecar-skip-webhook/1-install | Deployment:kuttl-test-willing-barnacle/vertx-create-span-sidecar created logger.go:42: 08:20:41 | sidecar-skip-webhook/1-install | test step completed 1-install logger.go:42: 08:20:41 | sidecar-skip-webhook/2-add-anotation-and-label | starting test step 2-add-anotation-and-label logger.go:42: 08:20:41 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name=jaeger-operator --namespace kuttl-test-willing-barnacle] logger.go:42: 08:20:41 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar labeled logger.go:42: 08:20:41 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-willing-barnacle] logger.go:42: 08:20:41 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 08:20:41 | sidecar-skip-webhook/2-add-anotation-and-label | test step completed 2-add-anotation-and-label logger.go:42: 08:20:41 | sidecar-skip-webhook/3-remove-label | starting test step 3-remove-label logger.go:42: 08:20:41 | sidecar-skip-webhook/3-remove-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name- --namespace kuttl-test-willing-barnacle] logger.go:42: 08:20:41 | sidecar-skip-webhook/3-remove-label | deployment.apps/vertx-create-span-sidecar unlabeled logger.go:42: 08:20:43 | sidecar-skip-webhook/3-remove-label | test step completed 3-remove-label logger.go:42: 08:20:43 | sidecar-skip-webhook | sidecar-skip-webhook events from ns kuttl-test-willing-barnacle: logger.go:42: 08:20:43 | sidecar-skip-webhook | 2024-05-20 08:20:36 +0000 UTC Normal Pod agent-as-sidecar-54784589b5-t7xh4 Binding Scheduled Successfully assigned kuttl-test-willing-barnacle/agent-as-sidecar-54784589b5-t7xh4 to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:20:43 | sidecar-skip-webhook | 2024-05-20 08:20:36 +0000 UTC Normal Pod agent-as-sidecar-54784589b5-t7xh4 AddedInterface Add eth0 [10.131.0.85/23] from ovn-kubernetes multus logger.go:42: 08:20:43 | sidecar-skip-webhook | 2024-05-20 08:20:36 +0000 UTC Normal Pod agent-as-sidecar-54784589b5-t7xh4.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" already present on machine kubelet logger.go:42: 08:20:43 | sidecar-skip-webhook | 2024-05-20 08:20:36 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-54784589b5 SuccessfulCreate Created pod: agent-as-sidecar-54784589b5-t7xh4 replicaset-controller logger.go:42: 08:20:43 | sidecar-skip-webhook | 2024-05-20 08:20:36 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-54784589b5 to 1 deployment-controller logger.go:42: 08:20:43 | sidecar-skip-webhook | 2024-05-20 08:20:37 +0000 UTC Normal Pod agent-as-sidecar-54784589b5-t7xh4.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:20:43 | sidecar-skip-webhook | 2024-05-20 08:20:37 +0000 UTC Normal Pod agent-as-sidecar-54784589b5-t7xh4.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:20:43 | sidecar-skip-webhook | 2024-05-20 08:20:39 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-fj9bl Binding Scheduled Successfully assigned kuttl-test-willing-barnacle/vertx-create-span-sidecar-84d458b68c-fj9bl to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:20:43 | sidecar-skip-webhook | 2024-05-20 08:20:39 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-fj9bl AddedInterface Add eth0 [10.129.2.49/23] from ovn-kubernetes multus logger.go:42: 08:20:43 | sidecar-skip-webhook | 2024-05-20 08:20:39 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-fj9bl.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 08:20:43 | sidecar-skip-webhook | 2024-05-20 08:20:39 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-fj9bl.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 08:20:43 | sidecar-skip-webhook | 2024-05-20 08:20:39 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-fj9bl.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 08:20:43 | sidecar-skip-webhook | 2024-05-20 08:20:39 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-fj9bl replicaset-controller logger.go:42: 08:20:43 | sidecar-skip-webhook | 2024-05-20 08:20:39 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 08:20:43 | sidecar-skip-webhook | 2024-05-20 08:20:41 +0000 UTC Normal Pod vertx-create-span-sidecar-6778fc9444-2vz7d Binding Scheduled Successfully assigned kuttl-test-willing-barnacle/vertx-create-span-sidecar-6778fc9444-2vz7d to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:20:43 | sidecar-skip-webhook | 2024-05-20 08:20:41 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-6778fc9444 SuccessfulCreate Created pod: vertx-create-span-sidecar-6778fc9444-2vz7d replicaset-controller logger.go:42: 08:20:43 | sidecar-skip-webhook | 2024-05-20 08:20:41 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-6778fc9444 to 1 deployment-controller logger.go:42: 08:20:43 | sidecar-skip-webhook | 2024-05-20 08:20:42 +0000 UTC Normal Pod vertx-create-span-sidecar-6778fc9444-2vz7d AddedInterface Add eth0 [10.128.2.68/23] from ovn-kubernetes multus logger.go:42: 08:20:43 | sidecar-skip-webhook | 2024-05-20 08:20:42 +0000 UTC Normal Pod vertx-create-span-sidecar-6778fc9444-2vz7d.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 08:20:43 | sidecar-skip-webhook | 2024-05-20 08:20:42 +0000 UTC Normal Pod vertx-create-span-sidecar-6778fc9444-2vz7d.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 08:20:43 | sidecar-skip-webhook | 2024-05-20 08:20:42 +0000 UTC Normal Pod vertx-create-span-sidecar-6778fc9444-2vz7d.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 08:20:43 | sidecar-skip-webhook | 2024-05-20 08:20:42 +0000 UTC Normal Pod vertx-create-span-sidecar-6778fc9444-2vz7d.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" already present on machine kubelet logger.go:42: 08:20:43 | sidecar-skip-webhook | 2024-05-20 08:20:42 +0000 UTC Normal Pod vertx-create-span-sidecar-6778fc9444-2vz7d.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:20:43 | sidecar-skip-webhook | 2024-05-20 08:20:42 +0000 UTC Normal Pod vertx-create-span-sidecar-6778fc9444-2vz7d.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:20:43 | sidecar-skip-webhook | Deleting namespace: kuttl-test-willing-barnacle === CONT kuttl/harness/sidecar-deployment logger.go:42: 08:20:49 | sidecar-deployment | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:20:49 | sidecar-deployment | Creating namespace: kuttl-test-clever-piglet logger.go:42: 08:20:50 | sidecar-deployment/0-install | starting test step 0-install logger.go:42: 08:20:50 | sidecar-deployment/0-install | Jaeger:kuttl-test-clever-piglet/agent-as-sidecar created logger.go:42: 08:20:56 | sidecar-deployment/0-install | test step completed 0-install logger.go:42: 08:20:56 | sidecar-deployment/1-install | starting test step 1-install logger.go:42: 08:20:56 | sidecar-deployment/1-install | Deployment:kuttl-test-clever-piglet/vertx-create-span-sidecar created logger.go:42: 08:20:58 | sidecar-deployment/1-install | test step completed 1-install logger.go:42: 08:20:58 | sidecar-deployment/2-enable-injection | starting test step 2-enable-injection logger.go:42: 08:20:58 | sidecar-deployment/2-enable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-clever-piglet] logger.go:42: 08:20:58 | sidecar-deployment/2-enable-injection | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 08:21:00 | sidecar-deployment/2-enable-injection | test step completed 2-enable-injection logger.go:42: 08:21:00 | sidecar-deployment/3-find-service | starting test step 3-find-service logger.go:42: 08:21:00 | sidecar-deployment/3-find-service | Job:kuttl-test-clever-piglet/00-find-service created logger.go:42: 08:21:12 | sidecar-deployment/3-find-service | test step completed 3-find-service logger.go:42: 08:21:12 | sidecar-deployment/4-other-instance | starting test step 4-other-instance logger.go:42: 08:21:12 | sidecar-deployment/4-other-instance | Jaeger:kuttl-test-clever-piglet/agent-as-sidecar2 created logger.go:42: 08:21:18 | sidecar-deployment/4-other-instance | test step completed 4-other-instance logger.go:42: 08:21:18 | sidecar-deployment/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 08:21:18 | sidecar-deployment/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 08:21:18 | sidecar-deployment/6-find-service | starting test step 6-find-service logger.go:42: 08:21:18 | sidecar-deployment/6-find-service | Job:kuttl-test-clever-piglet/01-find-service created logger.go:42: 08:21:39 | sidecar-deployment/6-find-service | test step completed 6-find-service logger.go:42: 08:21:39 | sidecar-deployment/7-disable-injection | starting test step 7-disable-injection logger.go:42: 08:21:39 | sidecar-deployment/7-disable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=false --namespace kuttl-test-clever-piglet] logger.go:42: 08:21:39 | sidecar-deployment/7-disable-injection | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 08:21:41 | sidecar-deployment/7-disable-injection | test step completed 7-disable-injection logger.go:42: 08:21:41 | sidecar-deployment | sidecar-deployment events from ns kuttl-test-clever-piglet: logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:20:53 +0000 UTC Normal Pod agent-as-sidecar-5f44684957-bsxvp Binding Scheduled Successfully assigned kuttl-test-clever-piglet/agent-as-sidecar-5f44684957-bsxvp to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:20:53 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-5f44684957 SuccessfulCreate Created pod: agent-as-sidecar-5f44684957-bsxvp replicaset-controller logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:20:53 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-5f44684957 to 1 deployment-controller logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:20:54 +0000 UTC Normal Pod agent-as-sidecar-5f44684957-bsxvp AddedInterface Add eth0 [10.131.0.86/23] from ovn-kubernetes multus logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:20:54 +0000 UTC Normal Pod agent-as-sidecar-5f44684957-bsxvp.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" already present on machine kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:20:54 +0000 UTC Normal Pod agent-as-sidecar-5f44684957-bsxvp.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:20:54 +0000 UTC Normal Pod agent-as-sidecar-5f44684957-bsxvp.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:20:56 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-68tjz Binding Scheduled Successfully assigned kuttl-test-clever-piglet/vertx-create-span-sidecar-84d458b68c-68tjz to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:20:56 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-68tjz replicaset-controller logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:20:56 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:20:57 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-68tjz AddedInterface Add eth0 [10.129.2.50/23] from ovn-kubernetes multus logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:20:57 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-68tjz.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:20:57 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-68tjz.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:20:57 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-68tjz.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:20:58 +0000 UTC Normal Pod vertx-create-span-sidecar-67f6bc97c8-7wkkp Binding Scheduled Successfully assigned kuttl-test-clever-piglet/vertx-create-span-sidecar-67f6bc97c8-7wkkp to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:20:58 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-67f6bc97c8 SuccessfulCreate Created pod: vertx-create-span-sidecar-67f6bc97c8-7wkkp replicaset-controller logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:20:58 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-67f6bc97c8 to 1 deployment-controller logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:20:59 +0000 UTC Normal Pod vertx-create-span-sidecar-67f6bc97c8-7wkkp AddedInterface Add eth0 [10.129.2.51/23] from ovn-kubernetes multus logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:20:59 +0000 UTC Normal Pod vertx-create-span-sidecar-67f6bc97c8-7wkkp.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:20:59 +0000 UTC Normal Pod vertx-create-span-sidecar-67f6bc97c8-7wkkp.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:20:59 +0000 UTC Normal Pod vertx-create-span-sidecar-67f6bc97c8-7wkkp.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:20:59 +0000 UTC Normal Pod vertx-create-span-sidecar-67f6bc97c8-7wkkp.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" already present on machine kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:20:59 +0000 UTC Normal Pod vertx-create-span-sidecar-67f6bc97c8-7wkkp.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:20:59 +0000 UTC Normal Pod vertx-create-span-sidecar-67f6bc97c8-7wkkp.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:00 +0000 UTC Normal Pod 00-find-service-fp8m6 Binding Scheduled Successfully assigned kuttl-test-clever-piglet/00-find-service-fp8m6 to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:00 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-fp8m6 job-controller logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:01 +0000 UTC Normal Pod 00-find-service-fp8m6 AddedInterface Add eth0 [10.128.2.69/23] from ovn-kubernetes multus logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:01 +0000 UTC Normal Pod 00-find-service-fp8m6.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:01 +0000 UTC Normal Pod 00-find-service-fp8m6.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:01 +0000 UTC Normal Pod 00-find-service-fp8m6.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:05 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-68tjz.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.50:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:05 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-68tjz.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.50:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:07 +0000 UTC Warning Pod vertx-create-span-sidecar-67f6bc97c8-7wkkp.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.51:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:07 +0000 UTC Warning Pod vertx-create-span-sidecar-67f6bc97c8-7wkkp.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.51:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:08 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-68tjz.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:08 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-68tjz.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.50:8080/": read tcp 10.129.2.2:60284->10.129.2.50:8080: read: connection reset by peer kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:08 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-68tjz.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.50:8080/": dial tcp 10.129.2.50:8080: connect: connection refused kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:09 +0000 UTC Normal Pod vertx-create-span-sidecar-67f6bc97c8-7wkkp.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:09 +0000 UTC Warning Pod vertx-create-span-sidecar-67f6bc97c8-7wkkp.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.51:8080/": read tcp 10.129.2.2:50490->10.129.2.51:8080: read: connection reset by peer kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:09 +0000 UTC Warning Pod vertx-create-span-sidecar-67f6bc97c8-7wkkp.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.51:8080/": dial tcp 10.129.2.51:8080: connect: connection refused kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:11 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:15 +0000 UTC Normal Pod agent-as-sidecar2-76dff6cd5-v6kfb Binding Scheduled Successfully assigned kuttl-test-clever-piglet/agent-as-sidecar2-76dff6cd5-v6kfb to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:15 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-76dff6cd5 SuccessfulCreate Created pod: agent-as-sidecar2-76dff6cd5-v6kfb replicaset-controller logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:15 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-76dff6cd5 to 1 deployment-controller logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:16 +0000 UTC Normal Pod agent-as-sidecar2-76dff6cd5-v6kfb AddedInterface Add eth0 [10.128.2.70/23] from ovn-kubernetes multus logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:16 +0000 UTC Normal Pod agent-as-sidecar2-76dff6cd5-v6kfb.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" already present on machine kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:16 +0000 UTC Normal Pod agent-as-sidecar2-76dff6cd5-v6kfb.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:16 +0000 UTC Normal Pod agent-as-sidecar2-76dff6cd5-v6kfb.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:18 +0000 UTC Normal Pod 01-find-service-ps7d4 Binding Scheduled Successfully assigned kuttl-test-clever-piglet/01-find-service-ps7d4 to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:18 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-ps7d4 job-controller logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:18 +0000 UTC Normal Pod agent-as-sidecar-5f44684957-bsxvp.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:19 +0000 UTC Normal Pod 01-find-service-ps7d4 AddedInterface Add eth0 [10.131.0.87/23] from ovn-kubernetes multus logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:19 +0000 UTC Normal Pod 01-find-service-ps7d4.spec.containers{asserts-container} Pulled Container image "registry.build05.ci.openshift.org/ci-op-m2wk24h7/pipeline@sha256:2bd8c55a5e94f12a139b30ca57a952bb429a1db60f8c49a4aa174359246d8337" already present on machine kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:19 +0000 UTC Normal Pod 01-find-service-ps7d4.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:19 +0000 UTC Normal Pod 01-find-service-ps7d4.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:19 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-68tjz.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.50:8080/": read tcp 10.129.2.2:52410->10.129.2.50:8080: read: connection reset by peer kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:22 +0000 UTC Normal Pod vertx-create-span-sidecar-7986c9b86d-jstt9 Binding Scheduled Successfully assigned kuttl-test-clever-piglet/vertx-create-span-sidecar-7986c9b86d-jstt9 to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:22 +0000 UTC Normal Pod vertx-create-span-sidecar-7986c9b86d-jstt9 AddedInterface Add eth0 [10.131.0.88/23] from ovn-kubernetes multus logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:22 +0000 UTC Normal Pod vertx-create-span-sidecar-7986c9b86d-jstt9.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:22 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-7986c9b86d SuccessfulCreate Created pod: vertx-create-span-sidecar-7986c9b86d-jstt9 replicaset-controller logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:22 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulDelete Deleted pod: vertx-create-span-sidecar-84d458b68c-68tjz replicaset-controller logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:22 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-84d458b68c to 0 from 1 deployment-controller logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:22 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-7986c9b86d to 1 from 0 deployment-controller logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:23 +0000 UTC Normal Pod vertx-create-span-sidecar-7986c9b86d-jstt9.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:23 +0000 UTC Normal Pod vertx-create-span-sidecar-7986c9b86d-jstt9.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:23 +0000 UTC Normal Pod vertx-create-span-sidecar-7986c9b86d-jstt9.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" already present on machine kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:23 +0000 UTC Normal Pod vertx-create-span-sidecar-7986c9b86d-jstt9.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:23 +0000 UTC Normal Pod vertx-create-span-sidecar-7986c9b86d-jstt9.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:31 +0000 UTC Warning Pod vertx-create-span-sidecar-7986c9b86d-jstt9.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.88:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:31 +0000 UTC Warning Pod vertx-create-span-sidecar-7986c9b86d-jstt9.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.131.0.88:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:33 +0000 UTC Normal Pod vertx-create-span-sidecar-7986c9b86d-jstt9.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:34 +0000 UTC Warning Pod vertx-create-span-sidecar-7986c9b86d-jstt9.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.88:8080/": read tcp 10.131.0.2:34228->10.131.0.88:8080: read: connection reset by peer kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:34 +0000 UTC Warning Pod vertx-create-span-sidecar-7986c9b86d-jstt9.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.88:8080/": dial tcp 10.131.0.88:8080: connect: connection refused kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:38 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:39 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-67f6bc97c8 SuccessfulDelete Deleted pod: vertx-create-span-sidecar-67f6bc97c8-7wkkp replicaset-controller logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:39 +0000 UTC Normal Pod vertx-create-span-sidecar-696d77d9cc-rlqgs Binding Scheduled Successfully assigned kuttl-test-clever-piglet/vertx-create-span-sidecar-696d77d9cc-rlqgs to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:39 +0000 UTC Normal Pod vertx-create-span-sidecar-696d77d9cc-rlqgs AddedInterface Add eth0 [10.128.2.71/23] from ovn-kubernetes multus logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:39 +0000 UTC Normal Pod vertx-create-span-sidecar-696d77d9cc-rlqgs.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:39 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-696d77d9cc SuccessfulCreate Created pod: vertx-create-span-sidecar-696d77d9cc-rlqgs replicaset-controller logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:39 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-67f6bc97c8 to 0 from 1 deployment-controller logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:39 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-696d77d9cc to 1 from 0 deployment-controller logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:40 +0000 UTC Normal Pod vertx-create-span-sidecar-696d77d9cc-rlqgs.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 08:21:41 | sidecar-deployment | 2024-05-20 08:21:40 +0000 UTC Normal Pod vertx-create-span-sidecar-696d77d9cc-rlqgs.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 08:21:41 | sidecar-deployment | Deleting namespace: kuttl-test-clever-piglet === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (148.74s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.92s) --- PASS: kuttl/harness/sidecar-namespace (66.77s) --- PASS: kuttl/harness/sidecar-skip-webhook (17.53s) --- PASS: kuttl/harness/sidecar-deployment (58.36s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name sidecar --report --output /logs/artifacts/sidecar.xml ./artifacts/kuttl-report.xml time="2024-05-20T08:21:48Z" level=debug msg="Setting a new name for the test suites" time="2024-05-20T08:21:48Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-05-20T08:21:48Z" level=debug msg="normalizing test case names" time="2024-05-20T08:21:48Z" level=debug msg="sidecar/artifacts -> sidecar_artifacts" time="2024-05-20T08:21:48Z" level=debug msg="sidecar/sidecar-namespace -> sidecar_sidecar_namespace" time="2024-05-20T08:21:48Z" level=debug msg="sidecar/sidecar-skip-webhook -> sidecar_sidecar_skip_webhook" time="2024-05-20T08:21:48Z" level=debug msg="sidecar/sidecar-deployment -> sidecar_sidecar_deployment" +------------------------------+--------+ | NAME | RESULT | +------------------------------+--------+ | sidecar_artifacts | passed | | sidecar_sidecar_namespace | passed | | sidecar_sidecar_skip_webhook | passed | | sidecar_sidecar_deployment | passed | +------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh streaming false true + '[' 3 -ne 3 ']' + test_suite_name=streaming + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/streaming.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-streaming make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ SKIP_ES_EXTERNAL=true \ ./tests/e2e/streaming/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-05-19-083311 True False 39m Cluster version is 4.16.0-0.nightly-2024-05-19-083311' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-05-19-083311 True False 39m Cluster version is 4.16.0-0.nightly-2024-05-19-083311' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/streaming/render.sh ++ export SUITE_DIR=./tests/e2e/streaming ++ SUITE_DIR=./tests/e2e/streaming ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/streaming ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + '[' false = true ']' + start_test streaming-simple + '[' 1 -ne 1 ']' + test_name=streaming-simple + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-simple' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-simple\e[0m' Rendering files for test streaming-simple + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build + '[' _build '!=' _build ']' + mkdir -p streaming-simple + cd streaming-simple + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/streaming-jaeger-assert.yaml.template -o ./04-assert.yaml + render_smoke_test simple-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=simple-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + export JAEGER_NAME=simple-streaming + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-tls + '[' 1 -ne 1 ']' + test_name=streaming-with-tls + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-tls' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-tls\e[0m' Rendering files for test streaming-with-tls + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-simple + '[' streaming-simple '!=' _build ']' + cd .. + mkdir -p streaming-with-tls + cd streaming-with-tls + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + render_smoke_test tls-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=tls-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + export JAEGER_NAME=tls-streaming + JAEGER_NAME=tls-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-autoprovisioning-autoscale + '[' 1 -ne 1 ']' + test_name=streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-autoprovisioning-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-autoprovisioning-autoscale\e[0m' Rendering files for test streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-with-tls + '[' streaming-with-tls '!=' _build ']' + cd .. + mkdir -p streaming-with-autoprovisioning-autoscale + cd streaming-with-autoprovisioning-autoscale + '[' true = true ']' + rm ./00-install.yaml ./00-assert.yaml + render_install_elasticsearch upstream 01 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=01 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./01-assert.yaml + jaeger_name=auto-provisioned + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="20Mi"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="500m"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.autoscale=true ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.minReplicas=1 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.maxReplicas=2 ./02-install.yaml + render_assert_kafka true auto-provisioned 03 + '[' 3 -ne 3 ']' + autoprovisioned=true + cluster_name=auto-provisioned + test_step=03 + '[' true = true ']' + is_kafka_minimal_enabled + namespaces=(observability openshift-operators openshift-distributed-tracing) + for i in "${namespaces[@]}" ++ kubectl get pods -n observability -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-operators -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-distributed-tracing -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled=true + '[' true == true ']' + return 0 + replicas=1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./03-assert.yaml ++ expr 03 + 1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./04-assert.yaml ++ expr 03 + 2 + CLUSTER_NAME=auto-provisioned + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./05-assert.yaml + version_lt 1.27 1.23 ++ echo 1.27 1.23 ++ tr ' ' '\n' ++ sort -rV ++ head -n 1 + test 1.27 '!=' 1.27 + rm ./08-assert.yaml + skip_test streaming-with-tls 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=streaming-with-tls + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-with-autoprovisioning-autoscale + '[' streaming-with-autoprovisioning-autoscale '!=' _build ']' + cd .. + rm -rf streaming-with-tls + warning 'streaming-with-tls: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: streaming-with-tls: This test is flaky in Prow CI\e[0m' WAR: streaming-with-tls: This test is flaky in Prow CI + skip_test streaming-simple 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=streaming-simple + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build + '[' _build '!=' _build ']' + rm -rf streaming-simple + warning 'streaming-simple: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: streaming-simple: This test is flaky in Prow CI\e[0m' WAR: streaming-simple: This test is flaky in Prow CI make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running streaming E2E tests' Running streaming E2E tests + cd tests/e2e/streaming/_build + set +e + KUBECONFIG=/tmp/kubeconfig-559878434 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 2 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/streaming-with-autoprovisioning-autoscale === PAUSE kuttl/harness/streaming-with-autoprovisioning-autoscale === CONT kuttl/harness/artifacts logger.go:42: 08:22:00 | artifacts | Creating namespace: kuttl-test-tidy-quetzal logger.go:42: 08:22:00 | artifacts | artifacts events from ns kuttl-test-tidy-quetzal: logger.go:42: 08:22:00 | artifacts | Deleting namespace: kuttl-test-tidy-quetzal === CONT kuttl/harness/streaming-with-autoprovisioning-autoscale logger.go:42: 08:22:06 | streaming-with-autoprovisioning-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:22:06 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_0.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:22:06 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_1.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:22:06 | streaming-with-autoprovisioning-autoscale | Creating namespace: kuttl-test-major-kid logger.go:42: 08:22:06 | streaming-with-autoprovisioning-autoscale/1-install | starting test step 1-install logger.go:42: 08:22:06 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc create sa deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 08:22:06 | streaming-with-autoprovisioning-autoscale/1-install | serviceaccount/deploy-elasticsearch created logger.go:42: 08:22:06 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc adm policy add-scc-to-user privileged -z deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 08:22:06 | streaming-with-autoprovisioning-autoscale/1-install | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:privileged added: "deploy-elasticsearch" logger.go:42: 08:22:06 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 6] logger.go:42: 08:22:12 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_0.yml -n $NAMESPACE] logger.go:42: 08:22:13 | streaming-with-autoprovisioning-autoscale/1-install | statefulset.apps/elasticsearch created logger.go:42: 08:22:13 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 3] logger.go:42: 08:22:16 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_1.yml -n $NAMESPACE] logger.go:42: 08:22:16 | streaming-with-autoprovisioning-autoscale/1-install | service/elasticsearch created logger.go:42: 08:22:34 | streaming-with-autoprovisioning-autoscale/1-install | test step completed 1-install logger.go:42: 08:22:34 | streaming-with-autoprovisioning-autoscale/2-install | starting test step 2-install logger.go:42: 08:22:34 | streaming-with-autoprovisioning-autoscale/2-install | Jaeger:kuttl-test-major-kid/auto-provisioned created logger.go:42: 08:22:34 | streaming-with-autoprovisioning-autoscale/2-install | test step completed 2-install logger.go:42: 08:22:34 | streaming-with-autoprovisioning-autoscale/3- | starting test step 3- logger.go:42: 08:23:25 | streaming-with-autoprovisioning-autoscale/3- | test step completed 3- logger.go:42: 08:23:25 | streaming-with-autoprovisioning-autoscale/4- | starting test step 4- logger.go:42: 08:23:57 | streaming-with-autoprovisioning-autoscale/4- | test step completed 4- logger.go:42: 08:23:57 | streaming-with-autoprovisioning-autoscale/5- | starting test step 5- logger.go:42: 08:24:30 | streaming-with-autoprovisioning-autoscale/5- | test step completed 5- logger.go:42: 08:24:30 | streaming-with-autoprovisioning-autoscale/6- | starting test step 6- logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale/6- | test step completed 6- logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale/7- | starting test step 7- logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale/7- | test step completed 7- logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | streaming-with-autoprovisioning-autoscale events from ns kuttl-test-major-kid: logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:22:13 +0000 UTC Normal Pod elasticsearch-0 Binding Scheduled Successfully assigned kuttl-test-major-kid/elasticsearch-0 to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:22:13 +0000 UTC Normal StatefulSet.apps elasticsearch SuccessfulCreate create Pod elasticsearch-0 in StatefulSet elasticsearch successful statefulset-controller logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:22:14 +0000 UTC Normal Pod elasticsearch-0 AddedInterface Add eth0 [10.131.0.89/23] from ovn-kubernetes multus logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:22:14 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulling Pulling image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:22:22 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulled Successfully pulled image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" in 8.036s (8.036s including waiting) kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:22:22 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:22:22 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:22:28 +0000 UTC Warning Pod elasticsearch-0.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Get "http://10.131.0.89:9200/": dial tcp 10.131.0.89:9200: connect: connection refused kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:22:39 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-zookeeper NoPods No matching pods found controllermanager logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:22:39 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:22:39 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:22:39 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-major-kid/data-auto-provisioned-zookeeper-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-7498f5b59f-rb6th_165985af-680d-4ea6-97ff-5ac0651c8094 logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:22:43 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ProvisioningSucceeded Successfully provisioned volume pvc-a9e436fc-dc70-487b-bafe-fcc5494177a1 ebs.csi.aws.com_aws-ebs-csi-driver-controller-7498f5b59f-rb6th_165985af-680d-4ea6-97ff-5ac0651c8094 logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:22:44 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 Binding Scheduled Successfully assigned kuttl-test-major-kid/auto-provisioned-zookeeper-0 to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:22:46 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-a9e436fc-dc70-487b-bafe-fcc5494177a1" attachdetach-controller logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:22:55 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 AddedInterface Add eth0 [10.129.2.52/23] from ovn-kubernetes multus logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:22:55 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulling Pulling image "registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:8b10bfb697b48ba3ed246fb846384f4cc67f05f670e9521edf0f47f829869404" kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:23:03 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulled Successfully pulled image "registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:8b10bfb697b48ba3ed246fb846384f4cc67f05f670e9521edf0f47f829869404" in 8.11s (8.11s including waiting) kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:23:03 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Created Created container zookeeper kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:23:03 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Started Started container zookeeper kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:23:27 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-kafka NoPods No matching pods found controllermanager logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:23:27 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:23:27 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-major-kid/data-0-auto-provisioned-kafka-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-7498f5b59f-rb6th_165985af-680d-4ea6-97ff-5ac0651c8094 logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:23:27 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:23:32 +0000 UTC Normal Pod auto-provisioned-kafka-0 Binding Scheduled Successfully assigned kuttl-test-major-kid/auto-provisioned-kafka-0 to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:23:32 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ProvisioningSucceeded Successfully provisioned volume pvc-c45f43b5-0c05-408d-bd52-94b09bf4eab1 ebs.csi.aws.com_aws-ebs-csi-driver-controller-7498f5b59f-rb6th_165985af-680d-4ea6-97ff-5ac0651c8094 logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:23:35 +0000 UTC Normal Pod auto-provisioned-kafka-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-c45f43b5-0c05-408d-bd52-94b09bf4eab1" attachdetach-controller logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:23:36 +0000 UTC Normal Pod auto-provisioned-kafka-0 AddedInterface Add eth0 [10.129.2.53/23] from ovn-kubernetes multus logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:23:36 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Pulled Container image "registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:8b10bfb697b48ba3ed246fb846384f4cc67f05f670e9521edf0f47f829869404" already present on machine kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:23:36 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Created Created container kafka kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:23:36 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Started Started container kafka kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:23:59 +0000 UTC Normal Pod auto-provisioned-entity-operator-78d8d7d499-jf4nc Binding Scheduled Successfully assigned kuttl-test-major-kid/auto-provisioned-entity-operator-78d8d7d499-jf4nc to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:23:59 +0000 UTC Normal Pod auto-provisioned-entity-operator-78d8d7d499-jf4nc AddedInterface Add eth0 [10.128.2.72/23] from ovn-kubernetes multus logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:23:59 +0000 UTC Normal Pod auto-provisioned-entity-operator-78d8d7d499-jf4nc.spec.containers{topic-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:dacf7ddffe7cbe2dc35e7d3290637b3334117e09dde986e6eaeaceded7069fbf" already present on machine kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:23:59 +0000 UTC Normal Pod auto-provisioned-entity-operator-78d8d7d499-jf4nc.spec.containers{topic-operator} Created Created container topic-operator kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:23:59 +0000 UTC Normal Pod auto-provisioned-entity-operator-78d8d7d499-jf4nc.spec.containers{topic-operator} Started Started container topic-operator kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:23:59 +0000 UTC Normal Pod auto-provisioned-entity-operator-78d8d7d499-jf4nc.spec.containers{user-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel8-operator@sha256:dacf7ddffe7cbe2dc35e7d3290637b3334117e09dde986e6eaeaceded7069fbf" already present on machine kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:23:59 +0000 UTC Normal Pod auto-provisioned-entity-operator-78d8d7d499-jf4nc.spec.containers{user-operator} Created Created container user-operator kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:23:59 +0000 UTC Normal Pod auto-provisioned-entity-operator-78d8d7d499-jf4nc.spec.containers{user-operator} Started Started container user-operator kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:23:59 +0000 UTC Normal Pod auto-provisioned-entity-operator-78d8d7d499-jf4nc.spec.containers{tls-sidecar} Pulling Pulling image "registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:8b10bfb697b48ba3ed246fb846384f4cc67f05f670e9521edf0f47f829869404" kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:23:59 +0000 UTC Normal ReplicaSet.apps auto-provisioned-entity-operator-78d8d7d499 SuccessfulCreate Created pod: auto-provisioned-entity-operator-78d8d7d499-jf4nc replicaset-controller logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:23:59 +0000 UTC Normal Deployment.apps auto-provisioned-entity-operator ScalingReplicaSet Scaled up replica set auto-provisioned-entity-operator-78d8d7d499 to 1 deployment-controller logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:08 +0000 UTC Normal Pod auto-provisioned-entity-operator-78d8d7d499-jf4nc.spec.containers{tls-sidecar} Pulled Successfully pulled image "registry.redhat.io/amq-streams/kafka-36-rhel8@sha256:8b10bfb697b48ba3ed246fb846384f4cc67f05f670e9521edf0f47f829869404" in 8.314s (8.314s including waiting) kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:08 +0000 UTC Normal Pod auto-provisioned-entity-operator-78d8d7d499-jf4nc.spec.containers{tls-sidecar} Created Created container tls-sidecar kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:08 +0000 UTC Normal Pod auto-provisioned-entity-operator-78d8d7d499-jf4nc.spec.containers{tls-sidecar} Started Started container tls-sidecar kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:32 +0000 UTC Normal Pod auto-provisioned-collector-695d4c59d9-qfvq6 Binding Scheduled Successfully assigned kuttl-test-major-kid/auto-provisioned-collector-695d4c59d9-qfvq6 to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:32 +0000 UTC Normal Pod auto-provisioned-collector-695d4c59d9-qfvq6 AddedInterface Add eth0 [10.131.0.90/23] from ovn-kubernetes multus logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:32 +0000 UTC Normal Pod auto-provisioned-collector-695d4c59d9-qfvq6.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:261a2c64c771323ad259e14802703bb790ee6b486b72c4939b487fbd83c30703" already present on machine kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:32 +0000 UTC Normal Pod auto-provisioned-collector-695d4c59d9-qfvq6.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:32 +0000 UTC Normal Pod auto-provisioned-collector-695d4c59d9-qfvq6.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:32 +0000 UTC Normal ReplicaSet.apps auto-provisioned-collector-695d4c59d9 SuccessfulCreate Created pod: auto-provisioned-collector-695d4c59d9-qfvq6 replicaset-controller logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:32 +0000 UTC Normal Deployment.apps auto-provisioned-collector ScalingReplicaSet Scaled up replica set auto-provisioned-collector-695d4c59d9 to 1 deployment-controller logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:32 +0000 UTC Normal Pod auto-provisioned-ingester-5588b7ffc5-cnmsf Binding Scheduled Successfully assigned kuttl-test-major-kid/auto-provisioned-ingester-5588b7ffc5-cnmsf to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:32 +0000 UTC Normal Pod auto-provisioned-ingester-5588b7ffc5-cnmsf AddedInterface Add eth0 [10.129.2.54/23] from ovn-kubernetes multus logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:32 +0000 UTC Normal Pod auto-provisioned-ingester-5588b7ffc5-cnmsf.spec.containers{jaeger-ingester} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:2072682c5abd59fb175478d506eaa37af363ad9244ce46c216aaf4c213113a24" kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:32 +0000 UTC Normal ReplicaSet.apps auto-provisioned-ingester-5588b7ffc5 SuccessfulCreate Created pod: auto-provisioned-ingester-5588b7ffc5-cnmsf replicaset-controller logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:32 +0000 UTC Normal Deployment.apps auto-provisioned-ingester ScalingReplicaSet Scaled up replica set auto-provisioned-ingester-5588b7ffc5 to 1 deployment-controller logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:32 +0000 UTC Normal Pod auto-provisioned-query-6bbb47f79d-f6zsg Binding Scheduled Successfully assigned kuttl-test-major-kid/auto-provisioned-query-6bbb47f79d-f6zsg to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:32 +0000 UTC Normal ReplicaSet.apps auto-provisioned-query-6bbb47f79d SuccessfulCreate Created pod: auto-provisioned-query-6bbb47f79d-f6zsg replicaset-controller logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:32 +0000 UTC Normal Deployment.apps auto-provisioned-query ScalingReplicaSet Scaled up replica set auto-provisioned-query-6bbb47f79d to 1 deployment-controller logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:33 +0000 UTC Warning Pod auto-provisioned-query-6bbb47f79d-f6zsg FailedMount MountVolume.SetUp failed for volume "auto-provisioned-service-ca" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:33 +0000 UTC Warning Pod auto-provisioned-query-6bbb47f79d-f6zsg FailedMount MountVolume.SetUp failed for volume "auto-provisioned-ui-oauth-proxy-tls" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:34 +0000 UTC Normal Pod auto-provisioned-query-6bbb47f79d-f6zsg AddedInterface Add eth0 [10.131.0.91/23] from ovn-kubernetes multus logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:34 +0000 UTC Normal Pod auto-provisioned-query-6bbb47f79d-f6zsg.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:156ea63cbd4dcd8c44173f118dba74388adbf1e82493237a4c514bb14131d5b8" already present on machine kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:34 +0000 UTC Normal Pod auto-provisioned-query-6bbb47f79d-f6zsg.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:34 +0000 UTC Normal Pod auto-provisioned-query-6bbb47f79d-f6zsg.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:34 +0000 UTC Normal Pod auto-provisioned-query-6bbb47f79d-f6zsg.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:34 +0000 UTC Normal Pod auto-provisioned-query-6bbb47f79d-f6zsg.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:34 +0000 UTC Normal Pod auto-provisioned-query-6bbb47f79d-f6zsg.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:34 +0000 UTC Normal Pod auto-provisioned-query-6bbb47f79d-f6zsg.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" already present on machine kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:34 +0000 UTC Normal Pod auto-provisioned-query-6bbb47f79d-f6zsg.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:34 +0000 UTC Normal Pod auto-provisioned-query-6bbb47f79d-f6zsg.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:39 +0000 UTC Normal Pod auto-provisioned-ingester-5588b7ffc5-cnmsf.spec.containers{jaeger-ingester} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:2072682c5abd59fb175478d506eaa37af363ad9244ce46c216aaf4c213113a24" in 6.731s (6.731s including waiting) kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:39 +0000 UTC Normal Pod auto-provisioned-ingester-5588b7ffc5-cnmsf.spec.containers{jaeger-ingester} Created Created container jaeger-ingester kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | 2024-05-20 08:24:39 +0000 UTC Normal Pod auto-provisioned-ingester-5588b7ffc5-cnmsf.spec.containers{jaeger-ingester} Started Started container jaeger-ingester kubelet logger.go:42: 08:24:41 | streaming-with-autoprovisioning-autoscale | Deleting namespace: kuttl-test-major-kid === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (201.46s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.90s) --- PASS: kuttl/harness/streaming-with-autoprovisioning-autoscale (195.41s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name streaming --report --output /logs/artifacts/streaming.xml ./artifacts/kuttl-report.xml time="2024-05-20T08:25:22Z" level=debug msg="Setting a new name for the test suites" time="2024-05-20T08:25:22Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-05-20T08:25:22Z" level=debug msg="normalizing test case names" time="2024-05-20T08:25:22Z" level=debug msg="streaming/artifacts -> streaming_artifacts" time="2024-05-20T08:25:22Z" level=debug msg="streaming/streaming-with-autoprovisioning-autoscale -> streaming_streaming_with_autoprovisioning_autoscale" +-----------------------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------------------+--------+ | streaming_artifacts | passed | | streaming_streaming_with_autoprovisioning_autoscale | passed | +-----------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh ui false true + '[' 3 -ne 3 ']' + test_suite_name=ui + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/ui.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-ui make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true ./tests/e2e/ui/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-05-19-083311 True False 43m Cluster version is 4.16.0-0.nightly-2024-05-19-083311' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.0-0.nightly-2024-05-19-083311 True False 43m Cluster version is 4.16.0-0.nightly-2024-05-19-083311' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/ui/render.sh ++ export SUITE_DIR=./tests/e2e/ui ++ SUITE_DIR=./tests/e2e/ui ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/ui ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test allinone + '[' 1 -ne 1 ']' + test_name=allinone + echo =========================================================================== =========================================================================== + info 'Rendering files for test allinone' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test allinone\e[0m' Rendering files for test allinone + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build + '[' _build '!=' _build ']' + mkdir -p allinone + cd allinone + export GET_URL_COMMAND + export URL + export JAEGER_NAME=all-in-one-ui + JAEGER_NAME=all-in-one-ui + '[' true = true ']' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./01-curl.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./04-test-ui-config.yaml + start_test production + '[' 1 -ne 1 ']' + test_name=production + echo =========================================================================== =========================================================================== + info 'Rendering files for test production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test production\e[0m' Rendering files for test production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build/allinone + '[' allinone '!=' _build ']' + cd .. + mkdir -p production + cd production + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + [[ true = true ]] + [[ true = true ]] + render_install_jaeger production-ui production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + '[' true = true ']' + INSECURE=true + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-forbbiden-access.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-curl.yaml + INSECURE=true + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./05-check-disabled-security.yaml + ASSERT_PRESENT=false + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./06-check-NO-gaID.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./08-check-gaID.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running ui E2E tests' Running ui E2E tests + cd tests/e2e/ui/_build + set +e + KUBECONFIG=/tmp/kubeconfig-559878434 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 3 tests === RUN kuttl/harness === RUN kuttl/harness/allinone === PAUSE kuttl/harness/allinone === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/production === PAUSE kuttl/harness/production === CONT kuttl/harness/allinone logger.go:42: 08:25:28 | allinone | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:25:28 | allinone | Creating namespace: kuttl-test-honest-chicken logger.go:42: 08:25:28 | allinone/0-install | starting test step 0-install logger.go:42: 08:25:29 | allinone/0-install | Jaeger:kuttl-test-honest-chicken/all-in-one-ui created logger.go:42: 08:25:33 | allinone/0-install | test step completed 0-install logger.go:42: 08:25:33 | allinone/1-curl | starting test step 1-curl logger.go:42: 08:25:33 | allinone/1-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 08:25:33 | allinone/1-curl | Checking the Ingress host value was populated logger.go:42: 08:25:33 | allinone/1-curl | Try number 0 logger.go:42: 08:25:33 | allinone/1-curl | Hostname is all-in-one-ui-kuttl-test-honest-chicken.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com logger.go:42: 08:25:33 | allinone/1-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE all-in-one-ui] logger.go:42: 08:25:33 | allinone/1-curl | Checking an expected HTTP response logger.go:42: 08:25:33 | allinone/1-curl | Running in OpenShift logger.go:42: 08:25:33 | allinone/1-curl | User not provided. Getting the token... logger.go:42: 08:25:35 | allinone/1-curl | Warning: resource jaegers/all-in-one-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 08:25:41 | allinone/1-curl | Try number 1/30 the https://all-in-one-ui-kuttl-test-honest-chicken.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 08:25:41 | allinone/1-curl | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 08:25:41 | allinone/1-curl | Try number 2/30 the https://all-in-one-ui-kuttl-test-honest-chicken.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 08:25:41 | allinone/1-curl | HTTP response is 503. 200 expected. Waiting 10 s logger.go:42: 08:25:51 | allinone/1-curl | Try number 3/30 the https://all-in-one-ui-kuttl-test-honest-chicken.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 08:25:51 | allinone/1-curl | curl response asserted properly logger.go:42: 08:25:51 | allinone/1-curl | test step completed 1-curl logger.go:42: 08:25:51 | allinone/2-delete | starting test step 2-delete logger.go:42: 08:25:51 | allinone/2-delete | Jaeger:kuttl-test-honest-chicken/all-in-one-ui created logger.go:42: 08:25:51 | allinone/2-delete | test step completed 2-delete logger.go:42: 08:25:51 | allinone/3-install | starting test step 3-install logger.go:42: 08:25:52 | allinone/3-install | Jaeger:kuttl-test-honest-chicken/all-in-one-ui updated logger.go:42: 08:25:52 | allinone/3-install | test step completed 3-install logger.go:42: 08:25:52 | allinone/4-test-ui-config | starting test step 4-test-ui-config logger.go:42: 08:25:52 | allinone/4-test-ui-config | running command: [./ensure-ingress-host.sh] logger.go:42: 08:25:52 | allinone/4-test-ui-config | Checking the Ingress host value was populated logger.go:42: 08:25:52 | allinone/4-test-ui-config | Try number 0 logger.go:42: 08:25:52 | allinone/4-test-ui-config | error: error executing jsonpath "{.items[0].status.ingress[0].host}": Error executing template: array index out of bounds: index 0, length 0. Printing more information for debugging the template: logger.go:42: 08:25:52 | allinone/4-test-ui-config | template was: logger.go:42: 08:25:52 | allinone/4-test-ui-config | {.items[0].status.ingress[0].host} logger.go:42: 08:25:52 | allinone/4-test-ui-config | object given to jsonpath engine was: logger.go:42: 08:25:52 | allinone/4-test-ui-config | map[string]interface {}{"apiVersion":"v1", "items":[]interface {}{}, "kind":"List", "metadata":map[string]interface {}{"resourceVersion":""}} logger.go:42: 08:25:52 | allinone/4-test-ui-config | logger.go:42: 08:25:52 | allinone/4-test-ui-config | logger.go:42: 08:26:02 | allinone/4-test-ui-config | Try number 1 logger.go:42: 08:26:02 | allinone/4-test-ui-config | Hostname is all-in-one-ui-kuttl-test-honest-chicken.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com logger.go:42: 08:26:02 | allinone/4-test-ui-config | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 08:26:02 | allinone/4-test-ui-config | time="2024-05-20T08:26:02Z" level=info msg="Querying https://all-in-one-ui-kuttl-test-honest-chicken.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com/search..." logger.go:42: 08:26:02 | allinone/4-test-ui-config | time="2024-05-20T08:26:02Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 08:26:02 | allinone/4-test-ui-config | time="2024-05-20T08:26:02Z" level=info msg="Polling to https://all-in-one-ui-kuttl-test-honest-chicken.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com/search" logger.go:42: 08:26:02 | allinone/4-test-ui-config | time="2024-05-20T08:26:02Z" level=info msg="Doing request number 0" logger.go:42: 08:26:02 | allinone/4-test-ui-config | time="2024-05-20T08:26:02Z" level=warning msg="Status code: 503" logger.go:42: 08:26:10 | allinone/4-test-ui-config | time="2024-05-20T08:26:10Z" level=info msg="Doing request number 1" logger.go:42: 08:26:10 | allinone/4-test-ui-config | time="2024-05-20T08:26:10Z" level=info msg="Content found and asserted!" logger.go:42: 08:26:10 | allinone/4-test-ui-config | time="2024-05-20T08:26:10Z" level=info msg="Success!" logger.go:42: 08:26:10 | allinone/4-test-ui-config | test step completed 4-test-ui-config logger.go:42: 08:26:10 | allinone | allinone events from ns kuttl-test-honest-chicken: logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:32 +0000 UTC Normal Pod all-in-one-ui-7f6f6864-9t4kz Binding Scheduled Successfully assigned kuttl-test-honest-chicken/all-in-one-ui-7f6f6864-9t4kz to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:32 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-7f6f6864 SuccessfulCreate Created pod: all-in-one-ui-7f6f6864-9t4kz replicaset-controller logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:32 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-7f6f6864 to 1 deployment-controller logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:33 +0000 UTC Normal Pod all-in-one-ui-7f6f6864-9t4kz AddedInterface Add eth0 [10.129.2.55/23] from ovn-kubernetes multus logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:33 +0000 UTC Normal Pod all-in-one-ui-7f6f6864-9t4kz.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" already present on machine kubelet logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:33 +0000 UTC Normal Pod all-in-one-ui-7f6f6864-9t4kz.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:33 +0000 UTC Normal Pod all-in-one-ui-7f6f6864-9t4kz.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:33 +0000 UTC Normal Pod all-in-one-ui-7f6f6864-9t4kz.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:33 +0000 UTC Normal Pod all-in-one-ui-7f6f6864-9t4kz.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:33 +0000 UTC Normal Pod all-in-one-ui-7f6f6864-9t4kz.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:37 +0000 UTC Normal Pod all-in-one-ui-7f6f6864-9t4kz.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:37 +0000 UTC Normal Pod all-in-one-ui-7f6f6864-9t4kz.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:37 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-7f6f6864 SuccessfulDelete Deleted pod: all-in-one-ui-7f6f6864-9t4kz replicaset-controller logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:37 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled down replica set all-in-one-ui-7f6f6864 to 0 from 1 deployment-controller logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:39 +0000 UTC Normal Pod all-in-one-ui-5d7d6b7466-xwvss Binding Scheduled Successfully assigned kuttl-test-honest-chicken/all-in-one-ui-5d7d6b7466-xwvss to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:39 +0000 UTC Normal Pod all-in-one-ui-5d7d6b7466-xwvss AddedInterface Add eth0 [10.129.2.56/23] from ovn-kubernetes multus logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:39 +0000 UTC Normal Pod all-in-one-ui-5d7d6b7466-xwvss.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" already present on machine kubelet logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:39 +0000 UTC Normal Pod all-in-one-ui-5d7d6b7466-xwvss.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:39 +0000 UTC Normal Pod all-in-one-ui-5d7d6b7466-xwvss.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:39 +0000 UTC Normal Pod all-in-one-ui-5d7d6b7466-xwvss.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:39 +0000 UTC Normal Pod all-in-one-ui-5d7d6b7466-xwvss.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:39 +0000 UTC Normal Pod all-in-one-ui-5d7d6b7466-xwvss.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:39 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-5d7d6b7466 SuccessfulCreate Created pod: all-in-one-ui-5d7d6b7466-xwvss replicaset-controller logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:39 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-5d7d6b7466 to 1 deployment-controller logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:52 +0000 UTC Normal Pod all-in-one-ui-5d7d6b7466-xwvss.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:52 +0000 UTC Normal Pod all-in-one-ui-5d7d6b7466-xwvss.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:57 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-74f98ff864 SuccessfulCreate Created pod: all-in-one-ui-74f98ff864-l27x5 replicaset-controller logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:57 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-74f98ff864 to 1 deployment-controller logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:58 +0000 UTC Normal Pod all-in-one-ui-74f98ff864-l27x5 Binding Scheduled Successfully assigned kuttl-test-honest-chicken/all-in-one-ui-74f98ff864-l27x5 to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:58 +0000 UTC Normal Pod all-in-one-ui-74f98ff864-l27x5 AddedInterface Add eth0 [10.129.2.57/23] from ovn-kubernetes multus logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:58 +0000 UTC Normal Pod all-in-one-ui-74f98ff864-l27x5.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:f077e804a0b4b791ab27d2dff643201b3fb6b3be59c1684b9576483d374b6fc3" already present on machine kubelet logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:58 +0000 UTC Normal Pod all-in-one-ui-74f98ff864-l27x5.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:26:10 | allinone | 2024-05-20 08:25:58 +0000 UTC Normal Pod all-in-one-ui-74f98ff864-l27x5.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:26:10 | allinone | Deleting namespace: kuttl-test-honest-chicken === CONT kuttl/harness/production logger.go:42: 08:26:17 | production | Ignoring add-tracking-id.yaml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:26:17 | production | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:26:17 | production | Creating namespace: kuttl-test-verified-racer logger.go:42: 08:26:17 | production/1-install | starting test step 1-install logger.go:42: 08:26:17 | production/1-install | Jaeger:kuttl-test-verified-racer/production-ui created logger.go:42: 08:26:55 | production/1-install | test step completed 1-install logger.go:42: 08:26:55 | production/2-check-forbbiden-access | starting test step 2-check-forbbiden-access logger.go:42: 08:26:55 | production/2-check-forbbiden-access | running command: [./ensure-ingress-host.sh] logger.go:42: 08:26:55 | production/2-check-forbbiden-access | Checking the Ingress host value was populated logger.go:42: 08:26:55 | production/2-check-forbbiden-access | Try number 0 logger.go:42: 08:26:55 | production/2-check-forbbiden-access | Hostname is production-ui-kuttl-test-verified-racer.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com logger.go:42: 08:26:55 | production/2-check-forbbiden-access | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE production-ui] logger.go:42: 08:26:56 | production/2-check-forbbiden-access | Checking an expected HTTP response logger.go:42: 08:26:56 | production/2-check-forbbiden-access | Running in OpenShift logger.go:42: 08:26:56 | production/2-check-forbbiden-access | Not using any secret logger.go:42: 08:26:56 | production/2-check-forbbiden-access | Try number 1/30 the https://production-ui-kuttl-test-verified-racer.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 08:26:56 | production/2-check-forbbiden-access | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 08:26:56 | production/2-check-forbbiden-access | Try number 2/30 the https://production-ui-kuttl-test-verified-racer.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 08:26:56 | production/2-check-forbbiden-access | curl response asserted properly logger.go:42: 08:26:56 | production/2-check-forbbiden-access | test step completed 2-check-forbbiden-access logger.go:42: 08:26:56 | production/3-curl | starting test step 3-curl logger.go:42: 08:26:56 | production/3-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 08:26:56 | production/3-curl | Checking the Ingress host value was populated logger.go:42: 08:26:56 | production/3-curl | Try number 0 logger.go:42: 08:26:56 | production/3-curl | Hostname is production-ui-kuttl-test-verified-racer.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com logger.go:42: 08:26:56 | production/3-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 08:26:56 | production/3-curl | Checking an expected HTTP response logger.go:42: 08:26:56 | production/3-curl | Running in OpenShift logger.go:42: 08:26:56 | production/3-curl | User not provided. Getting the token... logger.go:42: 08:26:58 | production/3-curl | Warning: resource jaegers/production-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 08:27:04 | production/3-curl | Try number 1/30 the https://production-ui-kuttl-test-verified-racer.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 08:27:04 | production/3-curl | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 08:27:04 | production/3-curl | Try number 2/30 the https://production-ui-kuttl-test-verified-racer.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 08:27:04 | production/3-curl | HTTP response is 503. 200 expected. Waiting 10 s logger.go:42: 08:27:14 | production/3-curl | Try number 3/30 the https://production-ui-kuttl-test-verified-racer.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 08:27:14 | production/3-curl | curl response asserted properly logger.go:42: 08:27:14 | production/3-curl | test step completed 3-curl logger.go:42: 08:27:14 | production/4-install | starting test step 4-install logger.go:42: 08:27:14 | production/4-install | Jaeger:kuttl-test-verified-racer/production-ui updated logger.go:42: 08:27:14 | production/4-install | test step completed 4-install logger.go:42: 08:27:14 | production/5-check-disabled-security | starting test step 5-check-disabled-security logger.go:42: 08:27:14 | production/5-check-disabled-security | running command: [./ensure-ingress-host.sh] logger.go:42: 08:27:14 | production/5-check-disabled-security | Checking the Ingress host value was populated logger.go:42: 08:27:14 | production/5-check-disabled-security | Try number 0 logger.go:42: 08:27:14 | production/5-check-disabled-security | Hostname is production-ui-kuttl-test-verified-racer.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com logger.go:42: 08:27:14 | production/5-check-disabled-security | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 08:27:15 | production/5-check-disabled-security | Checking an expected HTTP response logger.go:42: 08:27:15 | production/5-check-disabled-security | Running in OpenShift logger.go:42: 08:27:15 | production/5-check-disabled-security | Not using any secret logger.go:42: 08:27:15 | production/5-check-disabled-security | Try number 1/30 the https://production-ui-kuttl-test-verified-racer.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 08:27:15 | production/5-check-disabled-security | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 08:27:15 | production/5-check-disabled-security | Try number 2/30 the https://production-ui-kuttl-test-verified-racer.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 08:27:15 | production/5-check-disabled-security | HTTP response is 403. 200 expected. Waiting 10 s logger.go:42: 08:27:25 | production/5-check-disabled-security | Try number 3/30 the https://production-ui-kuttl-test-verified-racer.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com/search logger.go:42: 08:27:25 | production/5-check-disabled-security | curl response asserted properly logger.go:42: 08:27:25 | production/5-check-disabled-security | test step completed 5-check-disabled-security logger.go:42: 08:27:25 | production/6-check-NO-gaID | starting test step 6-check-NO-gaID logger.go:42: 08:27:25 | production/6-check-NO-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 08:27:25 | production/6-check-NO-gaID | Checking the Ingress host value was populated logger.go:42: 08:27:25 | production/6-check-NO-gaID | Try number 0 logger.go:42: 08:27:25 | production/6-check-NO-gaID | Hostname is production-ui-kuttl-test-verified-racer.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com logger.go:42: 08:27:25 | production/6-check-NO-gaID | running command: [sh -c ASSERT_PRESENT=false EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 08:27:25 | production/6-check-NO-gaID | time="2024-05-20T08:27:25Z" level=info msg="Querying https://production-ui-kuttl-test-verified-racer.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com/search..." logger.go:42: 08:27:25 | production/6-check-NO-gaID | time="2024-05-20T08:27:25Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 08:27:25 | production/6-check-NO-gaID | time="2024-05-20T08:27:25Z" level=info msg="Polling to https://production-ui-kuttl-test-verified-racer.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com/search" logger.go:42: 08:27:25 | production/6-check-NO-gaID | time="2024-05-20T08:27:25Z" level=info msg="Doing request number 0" logger.go:42: 08:27:25 | production/6-check-NO-gaID | time="2024-05-20T08:27:25Z" level=info msg="Content not found and asserted it was not found!" logger.go:42: 08:27:25 | production/6-check-NO-gaID | time="2024-05-20T08:27:25Z" level=info msg="Success!" logger.go:42: 08:27:25 | production/6-check-NO-gaID | test step completed 6-check-NO-gaID logger.go:42: 08:27:25 | production/7-add-tracking-id | starting test step 7-add-tracking-id logger.go:42: 08:27:25 | production/7-add-tracking-id | running command: [sh -c kubectl apply -f add-tracking-id.yaml -n $NAMESPACE] logger.go:42: 08:27:26 | production/7-add-tracking-id | jaeger.jaegertracing.io/production-ui configured logger.go:42: 08:27:26 | production/7-add-tracking-id | test step completed 7-add-tracking-id logger.go:42: 08:27:26 | production/8-check-gaID | starting test step 8-check-gaID logger.go:42: 08:27:26 | production/8-check-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 08:27:26 | production/8-check-gaID | Checking the Ingress host value was populated logger.go:42: 08:27:26 | production/8-check-gaID | Try number 0 logger.go:42: 08:27:26 | production/8-check-gaID | Hostname is production-ui-kuttl-test-verified-racer.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com logger.go:42: 08:27:26 | production/8-check-gaID | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 08:27:26 | production/8-check-gaID | time="2024-05-20T08:27:26Z" level=info msg="Querying https://production-ui-kuttl-test-verified-racer.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com/search..." logger.go:42: 08:27:26 | production/8-check-gaID | time="2024-05-20T08:27:26Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 08:27:26 | production/8-check-gaID | time="2024-05-20T08:27:26Z" level=info msg="Polling to https://production-ui-kuttl-test-verified-racer.apps.ci-op-m2wk24h7-c3652.cspilp.interop.ccitredhat.com/search" logger.go:42: 08:27:26 | production/8-check-gaID | time="2024-05-20T08:27:26Z" level=info msg="Doing request number 0" logger.go:42: 08:27:26 | production/8-check-gaID | time="2024-05-20T08:27:26Z" level=warning msg="Found: false . Assert: true" logger.go:42: 08:27:26 | production/8-check-gaID | time="2024-05-20T08:27:26Z" level=warning msg="The condition of the test function was not accomplished" logger.go:42: 08:27:34 | production/8-check-gaID | time="2024-05-20T08:27:34Z" level=info msg="Doing request number 1" logger.go:42: 08:27:54 | production/8-check-gaID | time="2024-05-20T08:27:54Z" level=info msg="Content found and asserted!" logger.go:42: 08:27:54 | production/8-check-gaID | time="2024-05-20T08:27:54Z" level=info msg="Success!" logger.go:42: 08:27:54 | production/8-check-gaID | test step completed 8-check-gaID logger.go:42: 08:27:54 | production | production events from ns kuttl-test-verified-racer: logger.go:42: 08:27:54 | production | 2024-05-20 08:26:23 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestverifiedracerproductionui-1-5846446fc6 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestverifiedracerproductionui-1-5846xxxk replicaset-controller logger.go:42: 08:27:54 | production | 2024-05-20 08:26:23 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestverifiedracerproductionui-1-5846xxxk Binding Scheduled Successfully assigned kuttl-test-verified-racer/elasticsearch-cdm-kuttltestverifiedracerproductionui-1-5846xxxk to ip-10-0-82-149.us-east-2.compute.internal default-scheduler logger.go:42: 08:27:54 | production | 2024-05-20 08:26:23 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestverifiedracerproductionui-1-5846xxxk AddedInterface Add eth0 [10.129.2.58/23] from ovn-kubernetes multus logger.go:42: 08:27:54 | production | 2024-05-20 08:26:23 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestverifiedracerproductionui-1-5846xxxk.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:32f336cb6d64ab7bce75fc5f7e2a01440400208bdffdbd965eec6823abcbd3f1" already present on machine kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:26:23 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestverifiedracerproductionui-1-5846xxxk.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:26:23 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestverifiedracerproductionui-1-5846xxxk.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:26:23 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestverifiedracerproductionui-1-5846xxxk.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:73d31fe7fb5609c1ff5b6edd3e746b4fbbba224a859ae7cf916c3af98c5c4ada" already present on machine kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:26:23 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestverifiedracerproductionui-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestverifiedracerproductionui-1-5846446fc6 to 1 deployment-controller logger.go:42: 08:27:54 | production | 2024-05-20 08:26:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestverifiedracerproductionui-1-5846xxxk.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:26:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestverifiedracerproductionui-1-5846xxxk.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:26:33 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestverifiedracerproductionui-1-5846xxxk.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:26:38 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestverifiedracerproductionui-1-5846xxxk.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:26:50 +0000 UTC Normal Pod production-ui-collector-69bc7c67d4-swb5m Binding Scheduled Successfully assigned kuttl-test-verified-racer/production-ui-collector-69bc7c67d4-swb5m to ip-10-0-8-197.us-east-2.compute.internal default-scheduler logger.go:42: 08:27:54 | production | 2024-05-20 08:26:50 +0000 UTC Normal ReplicaSet.apps production-ui-collector-69bc7c67d4 SuccessfulCreate Created pod: production-ui-collector-69bc7c67d4-swb5m replicaset-controller logger.go:42: 08:27:54 | production | 2024-05-20 08:26:50 +0000 UTC Normal Deployment.apps production-ui-collector ScalingReplicaSet Scaled up replica set production-ui-collector-69bc7c67d4 to 1 deployment-controller logger.go:42: 08:27:54 | production | 2024-05-20 08:26:50 +0000 UTC Normal Pod production-ui-query-6c55d9779f-jqhv7 Binding Scheduled Successfully assigned kuttl-test-verified-racer/production-ui-query-6c55d9779f-jqhv7 to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:27:54 | production | 2024-05-20 08:26:50 +0000 UTC Normal ReplicaSet.apps production-ui-query-6c55d9779f SuccessfulCreate Created pod: production-ui-query-6c55d9779f-jqhv7 replicaset-controller logger.go:42: 08:27:54 | production | 2024-05-20 08:26:50 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-6c55d9779f to 1 deployment-controller logger.go:42: 08:27:54 | production | 2024-05-20 08:26:51 +0000 UTC Warning Pod production-ui-collector-69bc7c67d4-swb5m FailedMount MountVolume.SetUp failed for volume "certs" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:26:51 +0000 UTC Warning Pod production-ui-collector-69bc7c67d4-swb5m FailedMount MountVolume.SetUp failed for volume "production-ui-sampling-configuration-volume" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:26:51 +0000 UTC Warning Pod production-ui-query-6c55d9779f-jqhv7 FailedMount MountVolume.SetUp failed for volume "production-ui-ui-configuration-volume" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:26:51 +0000 UTC Warning Pod production-ui-query-6c55d9779f-jqhv7 FailedMount MountVolume.SetUp failed for volume "production-ui-service-ca" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:26:51 +0000 UTC Warning Pod production-ui-query-6c55d9779f-jqhv7 FailedMount MountVolume.SetUp failed for volume "kube-api-access-qv2jz" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:26:52 +0000 UTC Normal Pod production-ui-collector-69bc7c67d4-swb5m AddedInterface Add eth0 [10.131.0.92/23] from ovn-kubernetes multus logger.go:42: 08:27:54 | production | 2024-05-20 08:26:52 +0000 UTC Normal Pod production-ui-collector-69bc7c67d4-swb5m.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:261a2c64c771323ad259e14802703bb790ee6b486b72c4939b487fbd83c30703" already present on machine kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:26:52 +0000 UTC Normal Pod production-ui-collector-69bc7c67d4-swb5m.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:26:52 +0000 UTC Normal Pod production-ui-collector-69bc7c67d4-swb5m.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:26:52 +0000 UTC Normal Pod production-ui-query-6c55d9779f-jqhv7 AddedInterface Add eth0 [10.128.2.73/23] from ovn-kubernetes multus logger.go:42: 08:27:54 | production | 2024-05-20 08:26:52 +0000 UTC Normal Pod production-ui-query-6c55d9779f-jqhv7.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:156ea63cbd4dcd8c44173f118dba74388adbf1e82493237a4c514bb14131d5b8" already present on machine kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:26:52 +0000 UTC Normal Pod production-ui-query-6c55d9779f-jqhv7.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:26:52 +0000 UTC Normal Pod production-ui-query-6c55d9779f-jqhv7.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:26:52 +0000 UTC Normal Pod production-ui-query-6c55d9779f-jqhv7.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:26:52 +0000 UTC Normal Pod production-ui-query-6c55d9779f-jqhv7.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:26:52 +0000 UTC Normal Pod production-ui-query-6c55d9779f-jqhv7.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:26:52 +0000 UTC Normal Pod production-ui-query-6c55d9779f-jqhv7.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" already present on machine kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:26:53 +0000 UTC Normal Pod production-ui-query-6c55d9779f-jqhv7.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:26:53 +0000 UTC Normal Pod production-ui-query-6c55d9779f-jqhv7.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:26:59 +0000 UTC Normal Pod production-ui-query-6c55d9779f-jqhv7.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:26:59 +0000 UTC Normal Pod production-ui-query-6c55d9779f-jqhv7.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:26:59 +0000 UTC Normal Pod production-ui-query-6c55d9779f-jqhv7.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:26:59 +0000 UTC Normal ReplicaSet.apps production-ui-query-6c55d9779f SuccessfulDelete Deleted pod: production-ui-query-6c55d9779f-jqhv7 replicaset-controller logger.go:42: 08:27:54 | production | 2024-05-20 08:26:59 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-6c55d9779f to 0 from 1 deployment-controller logger.go:42: 08:27:54 | production | 2024-05-20 08:27:00 +0000 UTC Normal Pod production-ui-query-747cf6d5d9-vzbh7 Binding Scheduled Successfully assigned kuttl-test-verified-racer/production-ui-query-747cf6d5d9-vzbh7 to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:27:54 | production | 2024-05-20 08:27:00 +0000 UTC Normal Pod production-ui-query-747cf6d5d9-vzbh7 AddedInterface Add eth0 [10.128.2.74/23] from ovn-kubernetes multus logger.go:42: 08:27:54 | production | 2024-05-20 08:27:00 +0000 UTC Normal Pod production-ui-query-747cf6d5d9-vzbh7.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:156ea63cbd4dcd8c44173f118dba74388adbf1e82493237a4c514bb14131d5b8" already present on machine kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:27:00 +0000 UTC Normal Pod production-ui-query-747cf6d5d9-vzbh7.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:27:00 +0000 UTC Normal Pod production-ui-query-747cf6d5d9-vzbh7.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:27:00 +0000 UTC Normal Pod production-ui-query-747cf6d5d9-vzbh7.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:65515661a7d719f1a8984c8fe3869a8c479c735b601d820c077797fe78b8e6b3" already present on machine kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:27:00 +0000 UTC Normal ReplicaSet.apps production-ui-query-747cf6d5d9 SuccessfulCreate Created pod: production-ui-query-747cf6d5d9-vzbh7 replicaset-controller logger.go:42: 08:27:54 | production | 2024-05-20 08:27:00 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-747cf6d5d9 to 1 deployment-controller logger.go:42: 08:27:54 | production | 2024-05-20 08:27:01 +0000 UTC Normal Pod production-ui-query-747cf6d5d9-vzbh7.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:27:01 +0000 UTC Normal Pod production-ui-query-747cf6d5d9-vzbh7.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:27:01 +0000 UTC Normal Pod production-ui-query-747cf6d5d9-vzbh7.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" already present on machine kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:27:01 +0000 UTC Normal Pod production-ui-query-747cf6d5d9-vzbh7.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:27:01 +0000 UTC Normal Pod production-ui-query-747cf6d5d9-vzbh7.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:27:05 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:27:54 | production | 2024-05-20 08:27:05 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:27:54 | production | 2024-05-20 08:27:05 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:27:54 | production | 2024-05-20 08:27:16 +0000 UTC Normal ReplicaSet.apps production-ui-query-5dc96d8c98 SuccessfulCreate Created pod: production-ui-query-5dc96d8c98-vk4v5 replicaset-controller logger.go:42: 08:27:54 | production | 2024-05-20 08:27:16 +0000 UTC Normal Pod production-ui-query-747cf6d5d9-vzbh7.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:27:16 +0000 UTC Normal Pod production-ui-query-747cf6d5d9-vzbh7.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:27:16 +0000 UTC Normal Pod production-ui-query-747cf6d5d9-vzbh7.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:27:16 +0000 UTC Normal ReplicaSet.apps production-ui-query-747cf6d5d9 SuccessfulDelete Deleted pod: production-ui-query-747cf6d5d9-vzbh7 replicaset-controller logger.go:42: 08:27:54 | production | 2024-05-20 08:27:16 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-747cf6d5d9 to 0 from 1 deployment-controller logger.go:42: 08:27:54 | production | 2024-05-20 08:27:16 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-5dc96d8c98 to 1 deployment-controller logger.go:42: 08:27:54 | production | 2024-05-20 08:27:17 +0000 UTC Normal Pod production-ui-query-5dc96d8c98-vk4v5 Binding Scheduled Successfully assigned kuttl-test-verified-racer/production-ui-query-5dc96d8c98-vk4v5 to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:27:54 | production | 2024-05-20 08:27:17 +0000 UTC Normal Pod production-ui-query-5dc96d8c98-vk4v5 AddedInterface Add eth0 [10.128.2.75/23] from ovn-kubernetes multus logger.go:42: 08:27:54 | production | 2024-05-20 08:27:17 +0000 UTC Normal Pod production-ui-query-5dc96d8c98-vk4v5.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:156ea63cbd4dcd8c44173f118dba74388adbf1e82493237a4c514bb14131d5b8" already present on machine kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:27:17 +0000 UTC Normal Pod production-ui-query-5dc96d8c98-vk4v5.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:27:17 +0000 UTC Normal Pod production-ui-query-5dc96d8c98-vk4v5.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:27:17 +0000 UTC Normal Pod production-ui-query-5dc96d8c98-vk4v5.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" already present on machine kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:27:17 +0000 UTC Normal Pod production-ui-query-5dc96d8c98-vk4v5.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:27:17 +0000 UTC Normal Pod production-ui-query-5dc96d8c98-vk4v5.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:27:27 +0000 UTC Normal Pod production-ui-query-5dc96d8c98-vk4v5.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:27:27 +0000 UTC Normal Pod production-ui-query-5dc96d8c98-vk4v5.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:27:27 +0000 UTC Normal ReplicaSet.apps production-ui-query-5dc96d8c98 SuccessfulDelete Deleted pod: production-ui-query-5dc96d8c98-vk4v5 replicaset-controller logger.go:42: 08:27:54 | production | 2024-05-20 08:27:27 +0000 UTC Normal Pod production-ui-query-c85ff7f49-x7mqj Binding Scheduled Successfully assigned kuttl-test-verified-racer/production-ui-query-c85ff7f49-x7mqj to ip-10-0-81-54.us-east-2.compute.internal default-scheduler logger.go:42: 08:27:54 | production | 2024-05-20 08:27:27 +0000 UTC Normal ReplicaSet.apps production-ui-query-c85ff7f49 SuccessfulCreate Created pod: production-ui-query-c85ff7f49-x7mqj replicaset-controller logger.go:42: 08:27:54 | production | 2024-05-20 08:27:27 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-5dc96d8c98 to 0 from 1 deployment-controller logger.go:42: 08:27:54 | production | 2024-05-20 08:27:27 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-c85ff7f49 to 1 deployment-controller logger.go:42: 08:27:54 | production | 2024-05-20 08:27:28 +0000 UTC Normal Pod production-ui-query-c85ff7f49-x7mqj AddedInterface Add eth0 [10.128.2.76/23] from ovn-kubernetes multus logger.go:42: 08:27:54 | production | 2024-05-20 08:27:28 +0000 UTC Normal Pod production-ui-query-c85ff7f49-x7mqj.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:156ea63cbd4dcd8c44173f118dba74388adbf1e82493237a4c514bb14131d5b8" already present on machine kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:27:29 +0000 UTC Normal Pod production-ui-query-c85ff7f49-x7mqj.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:27:29 +0000 UTC Normal Pod production-ui-query-c85ff7f49-x7mqj.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:27:29 +0000 UTC Normal Pod production-ui-query-c85ff7f49-x7mqj.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:dc88f8af5dca50238f8276518c9c6b50f9c886724cf690d27da073c93ee69976" already present on machine kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:27:29 +0000 UTC Normal Pod production-ui-query-c85ff7f49-x7mqj.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:27:29 +0000 UTC Normal Pod production-ui-query-c85ff7f49-x7mqj.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:27:54 | production | 2024-05-20 08:27:50 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod production-ui-collector-69bc7c67d4-swb5m horizontal-pod-autoscaler logger.go:42: 08:27:54 | production | 2024-05-20 08:27:50 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod production-ui-collector-69bc7c67d4-swb5m horizontal-pod-autoscaler logger.go:42: 08:27:54 | production | 2024-05-20 08:27:50 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod production-ui-collector-69bc7c67d4-swb5m horizontal-pod-autoscaler logger.go:42: 08:27:54 | production | Deleting namespace: kuttl-test-verified-racer === CONT kuttl/harness/artifacts logger.go:42: 08:28:01 | artifacts | Creating namespace: kuttl-test-select-man logger.go:42: 08:28:01 | artifacts | artifacts events from ns kuttl-test-select-man: logger.go:42: 08:28:01 | artifacts | Deleting namespace: kuttl-test-select-man === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (158.91s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/allinone (48.68s) --- PASS: kuttl/harness/production (104.00s) --- PASS: kuttl/harness/artifacts (6.07s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name ui --report --output /logs/artifacts/ui.xml ./artifacts/kuttl-report.xml time="2024-05-20T08:28:09Z" level=debug msg="Setting a new name for the test suites" time="2024-05-20T08:28:09Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-05-20T08:28:09Z" level=debug msg="normalizing test case names" time="2024-05-20T08:28:09Z" level=debug msg="ui/allinone -> ui_allinone" time="2024-05-20T08:28:09Z" level=debug msg="ui/production -> ui_production" time="2024-05-20T08:28:09Z" level=debug msg="ui/artifacts -> ui_artifacts" +---------------+--------+ | NAME | RESULT | +---------------+--------+ | ui_allinone | passed | | ui_production | passed | | ui_artifacts | passed | +---------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests'