% Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 3831 100 3831 0 0 22150 0 --:--:-- --:--:-- --:--:-- 22273 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 9315 100 9315 0 0 41998 0 --:--:-- --:--:-- --:--:-- 42149 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 2607 100 2607 0 0 18071 0 --:--:-- --:--:-- --:--:-- 18104 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 1953 100 1953 0 0 12102 0 --:--:-- --:--:-- --:--:-- 12130 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 350 100 350 0 0 2310 0 --:--:-- --:--:-- --:--:-- 2317 Installing kuttl Try 0... curl -sLo /tmp/jaeger-tests/hack/install/../../bin/kubectl-kuttl https://github.com/kudobuilder/kuttl/releases/download/v0.15.0/kubectl-kuttl_0.15.0_linux_x86_64 KUBECONFIG file is: /tmp/kubeconfig-3197734706 for suite in miscellaneous upgrade elasticsearch examples generate sidecar streaming ui; do \ make run-e2e-tests-$suite ; \ done make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh miscellaneous false true + '[' 3 -ne 3 ']' + test_suite_name=miscellaneous + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/miscellaneous.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-miscellaneous make[2]: Entering directory '/tmp/jaeger-tests' SKIP_ES_EXTERNAL=true ./tests/e2e/miscellaneous/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-15-030107 True False 10m Cluster version is 4.18.0-0.nightly-2025-01-15-030107' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-15-030107 True False 10m Cluster version is 4.18.0-0.nightly-2025-01-15-030107' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/miscellaneous/render.sh ++ export SUITE_DIR=./tests/e2e/miscellaneous ++ SUITE_DIR=./tests/e2e/miscellaneous ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/miscellaneous ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test collector-autoscale + '[' 1 -ne 1 ']' + test_name=collector-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-autoscale\e[0m' Rendering files for test collector-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p collector-autoscale + cd collector-autoscale + jaeger_name=simple-prod + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + ELASTICSEARCH_NODECOUNT=1 + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.resources.requests.memory="200m"' 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.autoscale=true 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.minReplicas=1 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.maxReplicas=2 01-install.yaml + version_lt 1.30 1.23 ++ echo 1.30 1.23 ++ tr ' ' '\n' ++ sort -rV ++ head -n 1 + test 1.30 '!=' 1.30 + rm ./03-assert.yaml + generate_otlp_e2e_tests http + test_protocol=http + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-http\e[0m' Rendering files for test collector-otlp-allinone-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-autoscale + '[' collector-autoscale '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-http + cd collector-otlp-allinone-http + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger http true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-http\e[0m' Rendering files for test collector-otlp-production-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-http + '[' collector-otlp-allinone-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-http + cd collector-otlp-production-http + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger http true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + generate_otlp_e2e_tests grpc + test_protocol=grpc + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-grpc\e[0m' Rendering files for test collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-http + '[' collector-otlp-production-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-grpc + cd collector-otlp-allinone-grpc + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-grpc\e[0m' Rendering files for test collector-otlp-production-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-grpc + '[' collector-otlp-allinone-grpc '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-grpc + cd collector-otlp-production-grpc + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + '[' true = true ']' + skip_test istio 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=istio + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-grpc + '[' collector-otlp-production-grpc '!=' _build ']' + cd .. + rm -rf istio + warning 'istio: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: istio: Test not supported in OpenShift\e[0m' WAR: istio: Test not supported in OpenShift + '[' true = true ']' + skip_test outside-cluster 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=outside-cluster + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + rm -rf outside-cluster + warning 'outside-cluster: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: outside-cluster: Test not supported in OpenShift\e[0m' WAR: outside-cluster: Test not supported in OpenShift + start_test set-custom-img + '[' 1 -ne 1 ']' + test_name=set-custom-img + echo =========================================================================== =========================================================================== + info 'Rendering files for test set-custom-img' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test set-custom-img\e[0m' Rendering files for test set-custom-img + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p set-custom-img + cd set-custom-img + jaeger_name=my-jaeger + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.image="test"' ./02-install.yaml + '[' true = true ']' + skip_test non-cluster-wide 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=non-cluster-wide + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/set-custom-img + '[' set-custom-img '!=' _build ']' + cd .. + rm -rf non-cluster-wide + warning 'non-cluster-wide: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: non-cluster-wide: Test not supported in OpenShift\e[0m' WAR: non-cluster-wide: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running miscellaneous E2E tests' Running miscellaneous E2E tests + cd tests/e2e/miscellaneous/_build + set +e + KUBECONFIG=/tmp/kubeconfig-3197734706 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 8 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/cassandra-spark === PAUSE kuttl/harness/cassandra-spark === RUN kuttl/harness/collector-autoscale === PAUSE kuttl/harness/collector-autoscale === RUN kuttl/harness/collector-otlp-allinone-grpc === PAUSE kuttl/harness/collector-otlp-allinone-grpc === RUN kuttl/harness/collector-otlp-allinone-http === PAUSE kuttl/harness/collector-otlp-allinone-http === RUN kuttl/harness/collector-otlp-production-grpc === PAUSE kuttl/harness/collector-otlp-production-grpc === RUN kuttl/harness/collector-otlp-production-http === PAUSE kuttl/harness/collector-otlp-production-http === RUN kuttl/harness/set-custom-img === PAUSE kuttl/harness/set-custom-img === CONT kuttl/harness/artifacts logger.go:42: 07:05:11 | artifacts | Creating namespace: kuttl-test-topical-molly logger.go:42: 07:05:11 | artifacts | artifacts events from ns kuttl-test-topical-molly: logger.go:42: 07:05:11 | artifacts | Deleting namespace: kuttl-test-topical-molly === CONT kuttl/harness/collector-otlp-allinone-http logger.go:42: 07:05:17 | collector-otlp-allinone-http | Creating namespace: kuttl-test-welcome-lacewing logger.go:42: 07:05:17 | collector-otlp-allinone-http/0-install | starting test step 0-install logger.go:42: 07:05:17 | collector-otlp-allinone-http/0-install | Jaeger:kuttl-test-welcome-lacewing/my-jaeger created logger.go:42: 07:05:32 | collector-otlp-allinone-http/0-install | test step completed 0-install logger.go:42: 07:05:32 | collector-otlp-allinone-http/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:05:32 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:05:34 | collector-otlp-allinone-http/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:05:42 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:05:43 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:05:43 | collector-otlp-allinone-http/1-smoke-test | job.batch/report-span created logger.go:42: 07:05:43 | collector-otlp-allinone-http/1-smoke-test | job.batch/check-span created logger.go:42: 07:06:01 | collector-otlp-allinone-http/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:06:02 | collector-otlp-allinone-http | collector-otlp-allinone-http events from ns kuttl-test-welcome-lacewing: logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:21 +0000 UTC Normal Pod my-jaeger-5b8d597d46-f8f92 Binding Scheduled Successfully assigned kuttl-test-welcome-lacewing/my-jaeger-5b8d597d46-f8f92 to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:21 +0000 UTC Warning Pod my-jaeger-5b8d597d46-f8f92 FailedMount MountVolume.SetUp failed for volume "my-jaeger-collector-tls-config-volume" : secret "my-jaeger-collector-headless-tls" not found kubelet logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:21 +0000 UTC Normal ReplicaSet.apps my-jaeger-5b8d597d46 SuccessfulCreate Created pod: my-jaeger-5b8d597d46-f8f92 replicaset-controller logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:21 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-5b8d597d46 to 1 deployment-controller logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:22 +0000 UTC Normal Pod my-jaeger-5b8d597d46-f8f92 AddedInterface Add eth0 [10.131.0.14/23] from ovn-kubernetes multus logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:22 +0000 UTC Normal Pod my-jaeger-5b8d597d46-f8f92.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" kubelet logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:28 +0000 UTC Normal Pod my-jaeger-5b8d597d46-f8f92.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" in 5.922s (5.922s including waiting). Image size: 147411949 bytes. kubelet logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:28 +0000 UTC Normal Pod my-jaeger-5b8d597d46-f8f92.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:28 +0000 UTC Normal Pod my-jaeger-5b8d597d46-f8f92.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:28 +0000 UTC Normal Pod my-jaeger-5b8d597d46-f8f92.spec.containers{oauth-proxy} Pulling Pulling image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" kubelet logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:31 +0000 UTC Normal Pod my-jaeger-5b8d597d46-f8f92.spec.containers{oauth-proxy} Pulled Successfully pulled image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" in 2.921s (2.921s including waiting). Image size: 339954870 bytes. kubelet logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:31 +0000 UTC Normal Pod my-jaeger-5b8d597d46-f8f92.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:31 +0000 UTC Normal Pod my-jaeger-5b8d597d46-f8f92.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:38 +0000 UTC Normal Pod my-jaeger-5b8d597d46-f8f92.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:38 +0000 UTC Normal Pod my-jaeger-5b8d597d46-f8f92.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:38 +0000 UTC Normal ReplicaSet.apps my-jaeger-5b8d597d46 SuccessfulDelete Deleted pod: my-jaeger-5b8d597d46-f8f92 replicaset-controller logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:38 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-5b8d597d46 to 0 from 1 deployment-controller logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:39 +0000 UTC Normal Pod my-jaeger-6769db554b-s6gjd Binding Scheduled Successfully assigned kuttl-test-welcome-lacewing/my-jaeger-6769db554b-s6gjd to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:39 +0000 UTC Normal Pod my-jaeger-6769db554b-s6gjd AddedInterface Add eth0 [10.131.0.15/23] from ovn-kubernetes multus logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:39 +0000 UTC Normal Pod my-jaeger-6769db554b-s6gjd.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:39 +0000 UTC Normal Pod my-jaeger-6769db554b-s6gjd.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:39 +0000 UTC Normal Pod my-jaeger-6769db554b-s6gjd.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:39 +0000 UTC Normal Pod my-jaeger-6769db554b-s6gjd.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:39 +0000 UTC Normal Pod my-jaeger-6769db554b-s6gjd.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:39 +0000 UTC Normal Pod my-jaeger-6769db554b-s6gjd.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:39 +0000 UTC Normal ReplicaSet.apps my-jaeger-6769db554b SuccessfulCreate Created pod: my-jaeger-6769db554b-s6gjd replicaset-controller logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:39 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-6769db554b to 1 deployment-controller logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:43 +0000 UTC Normal Pod check-span-96gdm Binding Scheduled Successfully assigned kuttl-test-welcome-lacewing/check-span-96gdm to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:43 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-96gdm job-controller logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:43 +0000 UTC Normal Pod report-span-fpmsb Binding Scheduled Successfully assigned kuttl-test-welcome-lacewing/report-span-fpmsb to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:43 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-fpmsb job-controller logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:44 +0000 UTC Normal Pod check-span-96gdm AddedInterface Add eth0 [10.128.2.26/23] from ovn-kubernetes multus logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:44 +0000 UTC Normal Pod check-span-96gdm.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:44 +0000 UTC Normal Pod report-span-fpmsb AddedInterface Add eth0 [10.129.2.18/23] from ovn-kubernetes multus logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:44 +0000 UTC Normal Pod report-span-fpmsb.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:46 +0000 UTC Normal Pod report-span-fpmsb.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 2.413s (2.413s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:46 +0000 UTC Normal Pod report-span-fpmsb.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:46 +0000 UTC Normal Pod report-span-fpmsb.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:58 +0000 UTC Normal Pod check-span-96gdm.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 14.045s (14.045s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:58 +0000 UTC Normal Pod check-span-96gdm.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:05:58 +0000 UTC Normal Pod check-span-96gdm.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:06:02 | collector-otlp-allinone-http | 2025-01-20 07:06:01 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:06:02 | collector-otlp-allinone-http | Deleting namespace: kuttl-test-welcome-lacewing === CONT kuttl/harness/set-custom-img logger.go:42: 07:06:14 | set-custom-img | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:06:14 | set-custom-img | Ignoring check-collector-img.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:06:14 | set-custom-img | Creating namespace: kuttl-test-holy-jaguar logger.go:42: 07:06:14 | set-custom-img/1-install | starting test step 1-install logger.go:42: 07:06:14 | set-custom-img/1-install | Jaeger:kuttl-test-holy-jaguar/my-jaeger created logger.go:42: 07:07:18 | set-custom-img/1-install | test step completed 1-install logger.go:42: 07:07:18 | set-custom-img/2-install | starting test step 2-install logger.go:42: 07:07:18 | set-custom-img/2-install | Jaeger:kuttl-test-holy-jaguar/my-jaeger updated logger.go:42: 07:07:18 | set-custom-img/2-install | test step completed 2-install logger.go:42: 07:07:18 | set-custom-img/3-check-image | starting test step 3-check-image logger.go:42: 07:07:18 | set-custom-img/3-check-image | running command: [sh -c ./check-collector-img.sh] logger.go:42: 07:07:18 | set-custom-img/3-check-image | Collector image missmatch. Expected: test. Has: registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d logger.go:42: 07:07:23 | set-custom-img/3-check-image | Collector image asserted properly! logger.go:42: 07:07:23 | set-custom-img/3-check-image | test step completed 3-check-image logger.go:42: 07:07:23 | set-custom-img | set-custom-img events from ns kuttl-test-holy-jaguar: logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:06:22 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestholyjaguarmyjaeger-1-5b9769ffd5 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestholyjaguarmyjaeger-1-5b9769ffd5nlbt5 replicaset-controller logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:06:22 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestholyjaguarmyjaeger-1-5b9769ffd5nlbt5 Binding Scheduled Successfully assigned kuttl-test-holy-jaguar/elasticsearch-cdm-kuttltestholyjaguarmyjaeger-1-5b9769ffd5nlbt5 to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:06:22 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestholyjaguarmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestholyjaguarmyjaeger-1-5b9769ffd5 to 1 deployment-controller logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:06:23 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestholyjaguarmyjaeger-1-5b9769ffd5nlbt5 AddedInterface Add eth0 [10.131.0.16/23] from ovn-kubernetes multus logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:06:23 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestholyjaguarmyjaeger-1-5b9769ffd5nlbt5.spec.containers{elasticsearch} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:06:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestholyjaguarmyjaeger-1-5b9769ffd5nlbt5.spec.containers{elasticsearch} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" in 9.064s (9.064s including waiting). Image size: 539381180 bytes. kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:06:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestholyjaguarmyjaeger-1-5b9769ffd5nlbt5.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:06:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestholyjaguarmyjaeger-1-5b9769ffd5nlbt5.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:06:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestholyjaguarmyjaeger-1-5b9769ffd5nlbt5.spec.containers{proxy} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:06:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestholyjaguarmyjaeger-1-5b9769ffd5nlbt5.spec.containers{proxy} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" in 3.076s (3.076s including waiting). Image size: 287327885 bytes. kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:06:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestholyjaguarmyjaeger-1-5b9769ffd5nlbt5.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:06:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestholyjaguarmyjaeger-1-5b9769ffd5nlbt5.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:06:43 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestholyjaguarmyjaeger-1-5b9769ffd5nlbt5.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:06:48 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestholyjaguarmyjaeger-1-5b9769ffd5nlbt5.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:06:59 +0000 UTC Normal Pod my-jaeger-collector-6d5bdd69ff-tfqp9 Binding Scheduled Successfully assigned kuttl-test-holy-jaguar/my-jaeger-collector-6d5bdd69ff-tfqp9 to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:06:59 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-6d5bdd69ff SuccessfulCreate Created pod: my-jaeger-collector-6d5bdd69ff-tfqp9 replicaset-controller logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:06:59 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-6d5bdd69ff to 1 deployment-controller logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:06:59 +0000 UTC Normal Pod my-jaeger-query-59454b8b88-ck898 Binding Scheduled Successfully assigned kuttl-test-holy-jaguar/my-jaeger-query-59454b8b88-ck898 to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:06:59 +0000 UTC Normal Pod my-jaeger-query-59454b8b88-ck898 AddedInterface Add eth0 [10.129.2.19/23] from ovn-kubernetes multus logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:06:59 +0000 UTC Normal Pod my-jaeger-query-59454b8b88-ck898.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:06:59 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-59454b8b88 SuccessfulCreate Created pod: my-jaeger-query-59454b8b88-ck898 replicaset-controller logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:06:59 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-59454b8b88 to 1 deployment-controller logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:00 +0000 UTC Warning Pod my-jaeger-collector-6d5bdd69ff-tfqp9 FailedMount MountVolume.SetUp failed for volume "my-jaeger-trusted-ca" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:00 +0000 UTC Warning Pod my-jaeger-collector-6d5bdd69ff-tfqp9 FailedMount MountVolume.SetUp failed for volume "my-jaeger-collector-tls-config-volume" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:01 +0000 UTC Normal Pod my-jaeger-collector-6d5bdd69ff-tfqp9 AddedInterface Add eth0 [10.128.2.27/23] from ovn-kubernetes multus logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:01 +0000 UTC Normal Pod my-jaeger-collector-6d5bdd69ff-tfqp9.spec.containers{jaeger-collector} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:06 +0000 UTC Normal Pod my-jaeger-query-59454b8b88-ck898.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" in 6.183s (6.183s including waiting). Image size: 192936692 bytes. kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:06 +0000 UTC Normal Pod my-jaeger-query-59454b8b88-ck898.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:06 +0000 UTC Normal Pod my-jaeger-query-59454b8b88-ck898.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:06 +0000 UTC Normal Pod my-jaeger-query-59454b8b88-ck898.spec.containers{oauth-proxy} Pulling Pulling image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:09 +0000 UTC Normal Pod my-jaeger-collector-6d5bdd69ff-tfqp9.spec.containers{jaeger-collector} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" in 8.306s (8.306s including waiting). Image size: 142489237 bytes. kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:09 +0000 UTC Normal Pod my-jaeger-collector-6d5bdd69ff-tfqp9.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:09 +0000 UTC Normal Pod my-jaeger-collector-6d5bdd69ff-tfqp9.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:09 +0000 UTC Normal Pod my-jaeger-query-59454b8b88-ck898.spec.containers{oauth-proxy} Pulled Successfully pulled image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" in 2.937s (2.937s including waiting). Image size: 339954870 bytes. kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:09 +0000 UTC Normal Pod my-jaeger-query-59454b8b88-ck898.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:09 +0000 UTC Normal Pod my-jaeger-query-59454b8b88-ck898.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:09 +0000 UTC Normal Pod my-jaeger-query-59454b8b88-ck898.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:14 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:14 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:14 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:17 +0000 UTC Normal Pod my-jaeger-query-59454b8b88-ck898.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" in 8.448s (8.448s including waiting). Image size: 115697219 bytes. kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:17 +0000 UTC Normal Pod my-jaeger-query-59454b8b88-ck898.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:17 +0000 UTC Normal Pod my-jaeger-query-59454b8b88-ck898.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:20 +0000 UTC Normal Pod my-jaeger-collector-6d5bdd69ff-tfqp9.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:20 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-6d5bdd69ff SuccessfulDelete Deleted pod: my-jaeger-collector-6d5bdd69ff-tfqp9 replicaset-controller logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:20 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-6d5bdd69ff to 0 from 1 deployment-controller logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:21 +0000 UTC Normal Pod my-jaeger-collector-6bcb4b4c58-jnrr2 Binding Scheduled Successfully assigned kuttl-test-holy-jaguar/my-jaeger-collector-6bcb4b4c58-jnrr2 to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:21 +0000 UTC Normal Pod my-jaeger-collector-6bcb4b4c58-jnrr2 AddedInterface Add eth0 [10.128.2.28/23] from ovn-kubernetes multus logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:21 +0000 UTC Normal Pod my-jaeger-collector-6bcb4b4c58-jnrr2.spec.containers{jaeger-collector} Pulling Pulling image "test" kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:21 +0000 UTC Warning Pod my-jaeger-collector-6bcb4b4c58-jnrr2.spec.containers{jaeger-collector} Failed Failed to pull image "test": initializing source docker://test:latest: reading manifest latest in docker.io/library/test: requested access to the resource is denied kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:21 +0000 UTC Warning Pod my-jaeger-collector-6bcb4b4c58-jnrr2.spec.containers{jaeger-collector} Failed Error: ErrImagePull kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:21 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-6bcb4b4c58 SuccessfulCreate Created pod: my-jaeger-collector-6bcb4b4c58-jnrr2 replicaset-controller logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:21 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-6bcb4b4c58 to 1 deployment-controller logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:22 +0000 UTC Normal Pod my-jaeger-collector-6bcb4b4c58-jnrr2.spec.containers{jaeger-collector} BackOff Back-off pulling image "test" kubelet logger.go:42: 07:07:23 | set-custom-img | 2025-01-20 07:07:22 +0000 UTC Warning Pod my-jaeger-collector-6bcb4b4c58-jnrr2.spec.containers{jaeger-collector} Failed Error: ImagePullBackOff kubelet logger.go:42: 07:07:23 | set-custom-img | Deleting namespace: kuttl-test-holy-jaguar === CONT kuttl/harness/collector-otlp-production-http logger.go:42: 07:07:30 | collector-otlp-production-http | Creating namespace: kuttl-test-divine-goat logger.go:42: 07:07:30 | collector-otlp-production-http/1-install | starting test step 1-install logger.go:42: 07:07:30 | collector-otlp-production-http/1-install | Jaeger:kuttl-test-divine-goat/my-jaeger created logger.go:42: 07:08:08 | collector-otlp-production-http/1-install | test step completed 1-install logger.go:42: 07:08:08 | collector-otlp-production-http/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:08:08 | collector-otlp-production-http/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:08:09 | collector-otlp-production-http/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:08:17 | collector-otlp-production-http/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:08:18 | collector-otlp-production-http/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:08:18 | collector-otlp-production-http/2-smoke-test | job.batch/report-span created logger.go:42: 07:08:18 | collector-otlp-production-http/2-smoke-test | job.batch/check-span created logger.go:42: 07:08:30 | collector-otlp-production-http/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:08:30 | collector-otlp-production-http | collector-otlp-production-http events from ns kuttl-test-divine-goat: logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:07:35 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestdivinegoatmyjaeger-1-69fd5f56df SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestdivinegoatmyjaeger-1-69fd5f56df9lwz6 replicaset-controller logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:07:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdivinegoatmyjaeger-1-69fd5f56df9lwz6 Binding Scheduled Successfully assigned kuttl-test-divine-goat/elasticsearch-cdm-kuttltestdivinegoatmyjaeger-1-69fd5f56df9lwz6 to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:07:35 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestdivinegoatmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestdivinegoatmyjaeger-1-69fd5f56df to 1 deployment-controller logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:07:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdivinegoatmyjaeger-1-69fd5f56df9lwz6 AddedInterface Add eth0 [10.131.0.17/23] from ovn-kubernetes multus logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:07:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdivinegoatmyjaeger-1-69fd5f56df9lwz6.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:07:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdivinegoatmyjaeger-1-69fd5f56df9lwz6.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:07:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdivinegoatmyjaeger-1-69fd5f56df9lwz6.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:07:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdivinegoatmyjaeger-1-69fd5f56df9lwz6.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:07:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdivinegoatmyjaeger-1-69fd5f56df9lwz6.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:07:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdivinegoatmyjaeger-1-69fd5f56df9lwz6.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:07:46 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestdivinegoatmyjaeger-1-69fd5f56df9lwz6.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:07:51 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestdivinegoatmyjaeger-1-69fd5f56df9lwz6.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:02 +0000 UTC Normal Pod my-jaeger-collector-57868c9bcc-hjq7v Binding Scheduled Successfully assigned kuttl-test-divine-goat/my-jaeger-collector-57868c9bcc-hjq7v to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:02 +0000 UTC Warning Pod my-jaeger-collector-57868c9bcc-hjq7v FailedMount MountVolume.SetUp failed for volume "my-jaeger-collector-tls-config-volume" : secret "my-jaeger-collector-headless-tls" not found kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:02 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-57868c9bcc SuccessfulCreate Created pod: my-jaeger-collector-57868c9bcc-hjq7v replicaset-controller logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:02 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-57868c9bcc to 1 deployment-controller logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:02 +0000 UTC Normal Pod my-jaeger-query-dc59d7b99-ghjxs Binding Scheduled Successfully assigned kuttl-test-divine-goat/my-jaeger-query-dc59d7b99-ghjxs to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:02 +0000 UTC Warning Pod my-jaeger-query-dc59d7b99-ghjxs FailedMount MountVolume.SetUp failed for volume "my-jaeger-ui-oauth-proxy-tls" : secret "my-jaeger-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:02 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-dc59d7b99 SuccessfulCreate Created pod: my-jaeger-query-dc59d7b99-ghjxs replicaset-controller logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:02 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-dc59d7b99 to 1 deployment-controller logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:03 +0000 UTC Normal Pod my-jaeger-collector-57868c9bcc-hjq7v AddedInterface Add eth0 [10.129.2.20/23] from ovn-kubernetes multus logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:03 +0000 UTC Normal Pod my-jaeger-collector-57868c9bcc-hjq7v.spec.containers{jaeger-collector} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:03 +0000 UTC Normal Pod my-jaeger-query-dc59d7b99-ghjxs AddedInterface Add eth0 [10.129.2.21/23] from ovn-kubernetes multus logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:03 +0000 UTC Normal Pod my-jaeger-query-dc59d7b99-ghjxs.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:04 +0000 UTC Normal Pod my-jaeger-query-dc59d7b99-ghjxs.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:04 +0000 UTC Normal Pod my-jaeger-query-dc59d7b99-ghjxs.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:04 +0000 UTC Normal Pod my-jaeger-query-dc59d7b99-ghjxs.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:04 +0000 UTC Normal Pod my-jaeger-query-dc59d7b99-ghjxs.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:04 +0000 UTC Normal Pod my-jaeger-query-dc59d7b99-ghjxs.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:04 +0000 UTC Normal Pod my-jaeger-query-dc59d7b99-ghjxs.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:04 +0000 UTC Normal Pod my-jaeger-query-dc59d7b99-ghjxs.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:04 +0000 UTC Normal Pod my-jaeger-query-dc59d7b99-ghjxs.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:06 +0000 UTC Normal Pod my-jaeger-collector-57868c9bcc-hjq7v.spec.containers{jaeger-collector} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" in 2.326s (2.326s including waiting). Image size: 142489237 bytes. kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:06 +0000 UTC Normal Pod my-jaeger-collector-57868c9bcc-hjq7v.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:06 +0000 UTC Normal Pod my-jaeger-collector-57868c9bcc-hjq7v.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:12 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-dc59d7b99 to 0 from 1 deployment-controller logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:13 +0000 UTC Normal Pod my-jaeger-query-859455f784-bn8x8 Binding Scheduled Successfully assigned kuttl-test-divine-goat/my-jaeger-query-859455f784-bn8x8 to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:13 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-859455f784 SuccessfulCreate Created pod: my-jaeger-query-859455f784-bn8x8 replicaset-controller logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:13 +0000 UTC Normal Pod my-jaeger-query-dc59d7b99-ghjxs.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:13 +0000 UTC Normal Pod my-jaeger-query-dc59d7b99-ghjxs.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:13 +0000 UTC Normal Pod my-jaeger-query-dc59d7b99-ghjxs.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:13 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-dc59d7b99 SuccessfulDelete Deleted pod: my-jaeger-query-dc59d7b99-ghjxs replicaset-controller logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:13 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-859455f784 to 1 deployment-controller logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:14 +0000 UTC Normal Pod my-jaeger-query-859455f784-bn8x8 AddedInterface Add eth0 [10.129.2.22/23] from ovn-kubernetes multus logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:14 +0000 UTC Normal Pod my-jaeger-query-859455f784-bn8x8.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:14 +0000 UTC Normal Pod my-jaeger-query-859455f784-bn8x8.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:14 +0000 UTC Normal Pod my-jaeger-query-859455f784-bn8x8.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:14 +0000 UTC Normal Pod my-jaeger-query-859455f784-bn8x8.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:14 +0000 UTC Normal Pod my-jaeger-query-859455f784-bn8x8.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:14 +0000 UTC Normal Pod my-jaeger-query-859455f784-bn8x8.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:14 +0000 UTC Normal Pod my-jaeger-query-859455f784-bn8x8.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:14 +0000 UTC Normal Pod my-jaeger-query-859455f784-bn8x8.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:14 +0000 UTC Normal Pod my-jaeger-query-859455f784-bn8x8.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:17 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:17 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:17 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:18 +0000 UTC Normal Pod check-span-8p66n Binding Scheduled Successfully assigned kuttl-test-divine-goat/check-span-8p66n to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:18 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-8p66n job-controller logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:18 +0000 UTC Normal Pod report-span-trrph Binding Scheduled Successfully assigned kuttl-test-divine-goat/report-span-trrph to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:18 +0000 UTC Normal Pod report-span-trrph AddedInterface Add eth0 [10.128.2.29/23] from ovn-kubernetes multus logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:18 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-trrph job-controller logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:19 +0000 UTC Normal Pod check-span-8p66n AddedInterface Add eth0 [10.128.2.30/23] from ovn-kubernetes multus logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:19 +0000 UTC Normal Pod check-span-8p66n.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:19 +0000 UTC Normal Pod report-span-trrph.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:19 +0000 UTC Normal Pod report-span-trrph.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 585ms (585ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:19 +0000 UTC Normal Pod report-span-trrph.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:19 +0000 UTC Normal Pod report-span-trrph.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:20 +0000 UTC Normal Pod check-span-8p66n.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.075s (1.075s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:20 +0000 UTC Normal Pod check-span-8p66n.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:20 +0000 UTC Normal Pod check-span-8p66n.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:08:30 | collector-otlp-production-http | 2025-01-20 07:08:30 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:08:30 | collector-otlp-production-http | Deleting namespace: kuttl-test-divine-goat === CONT kuttl/harness/collector-otlp-production-grpc logger.go:42: 07:08:38 | collector-otlp-production-grpc | Creating namespace: kuttl-test-thorough-reindeer logger.go:42: 07:08:38 | collector-otlp-production-grpc/1-install | starting test step 1-install logger.go:42: 07:08:38 | collector-otlp-production-grpc/1-install | Jaeger:kuttl-test-thorough-reindeer/my-jaeger created logger.go:42: 07:09:15 | collector-otlp-production-grpc/1-install | test step completed 1-install logger.go:42: 07:09:15 | collector-otlp-production-grpc/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:09:15 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:09:16 | collector-otlp-production-grpc/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:09:22 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:09:23 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:09:23 | collector-otlp-production-grpc/2-smoke-test | job.batch/report-span created logger.go:42: 07:09:23 | collector-otlp-production-grpc/2-smoke-test | job.batch/check-span created logger.go:42: 07:09:45 | collector-otlp-production-grpc/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:09:45 | collector-otlp-production-grpc | collector-otlp-production-grpc events from ns kuttl-test-thorough-reindeer: logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:08:45 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestthoroughreindeermyjaeger-1-5c4bd74957 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestthoroughreindeermyjaeger-1-5c4bkr67q replicaset-controller logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:08:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestthoroughreindeermyjaeger-1-5c4bkr67q Binding Scheduled Successfully assigned kuttl-test-thorough-reindeer/elasticsearch-cdm-kuttltestthoroughreindeermyjaeger-1-5c4bkr67q to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:08:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestthoroughreindeermyjaeger-1-5c4bkr67q AddedInterface Add eth0 [10.131.0.18/23] from ovn-kubernetes multus logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:08:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestthoroughreindeermyjaeger-1-5c4bkr67q.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:08:45 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestthoroughreindeermyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestthoroughreindeermyjaeger-1-5c4bd74957 to 1 deployment-controller logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:08:46 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestthoroughreindeermyjaeger-1-5c4bkr67q.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:08:46 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestthoroughreindeermyjaeger-1-5c4bkr67q.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:08:46 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestthoroughreindeermyjaeger-1-5c4bkr67q.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:08:46 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestthoroughreindeermyjaeger-1-5c4bkr67q.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:08:46 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestthoroughreindeermyjaeger-1-5c4bkr67q.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:00 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestthoroughreindeermyjaeger-1-5c4bkr67q.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:12 +0000 UTC Normal Pod my-jaeger-collector-75ccc45495-dhwtk Binding Scheduled Successfully assigned kuttl-test-thorough-reindeer/my-jaeger-collector-75ccc45495-dhwtk to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:12 +0000 UTC Normal Pod my-jaeger-collector-75ccc45495-dhwtk AddedInterface Add eth0 [10.129.2.23/23] from ovn-kubernetes multus logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:12 +0000 UTC Normal Pod my-jaeger-collector-75ccc45495-dhwtk.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:12 +0000 UTC Normal Pod my-jaeger-collector-75ccc45495-dhwtk.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:12 +0000 UTC Normal Pod my-jaeger-collector-75ccc45495-dhwtk.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:12 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-75ccc45495 SuccessfulCreate Created pod: my-jaeger-collector-75ccc45495-dhwtk replicaset-controller logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:12 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-75ccc45495 to 1 deployment-controller logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:12 +0000 UTC Normal Pod my-jaeger-query-89657ffd4-2zj94 Binding Scheduled Successfully assigned kuttl-test-thorough-reindeer/my-jaeger-query-89657ffd4-2zj94 to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:12 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-89657ffd4 SuccessfulCreate Created pod: my-jaeger-query-89657ffd4-2zj94 replicaset-controller logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:12 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-89657ffd4 to 1 deployment-controller logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:13 +0000 UTC Normal Pod my-jaeger-query-89657ffd4-2zj94 AddedInterface Add eth0 [10.129.2.24/23] from ovn-kubernetes multus logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:13 +0000 UTC Normal Pod my-jaeger-query-89657ffd4-2zj94.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:13 +0000 UTC Normal Pod my-jaeger-query-89657ffd4-2zj94.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:13 +0000 UTC Normal Pod my-jaeger-query-89657ffd4-2zj94.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:13 +0000 UTC Normal Pod my-jaeger-query-89657ffd4-2zj94.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:13 +0000 UTC Normal Pod my-jaeger-query-89657ffd4-2zj94.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:13 +0000 UTC Normal Pod my-jaeger-query-89657ffd4-2zj94.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:13 +0000 UTC Normal Pod my-jaeger-query-89657ffd4-2zj94.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:13 +0000 UTC Normal Pod my-jaeger-query-89657ffd4-2zj94.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:13 +0000 UTC Normal Pod my-jaeger-query-89657ffd4-2zj94.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:19 +0000 UTC Normal Pod my-jaeger-query-89657ffd4-2zj94.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:19 +0000 UTC Normal Pod my-jaeger-query-89657ffd4-2zj94.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:19 +0000 UTC Normal Pod my-jaeger-query-89657ffd4-2zj94.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:19 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-89657ffd4 SuccessfulDelete Deleted pod: my-jaeger-query-89657ffd4-2zj94 replicaset-controller logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:19 +0000 UTC Normal Pod my-jaeger-query-bcf5bf78f-ltxjm Binding Scheduled Successfully assigned kuttl-test-thorough-reindeer/my-jaeger-query-bcf5bf78f-ltxjm to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:19 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-bcf5bf78f SuccessfulCreate Created pod: my-jaeger-query-bcf5bf78f-ltxjm replicaset-controller logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:19 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-89657ffd4 to 0 from 1 deployment-controller logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:19 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-bcf5bf78f to 1 deployment-controller logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:20 +0000 UTC Normal Pod my-jaeger-query-bcf5bf78f-ltxjm AddedInterface Add eth0 [10.129.2.25/23] from ovn-kubernetes multus logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:20 +0000 UTC Normal Pod my-jaeger-query-bcf5bf78f-ltxjm.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:20 +0000 UTC Normal Pod my-jaeger-query-bcf5bf78f-ltxjm.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:20 +0000 UTC Normal Pod my-jaeger-query-bcf5bf78f-ltxjm.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:20 +0000 UTC Normal Pod my-jaeger-query-bcf5bf78f-ltxjm.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:20 +0000 UTC Normal Pod my-jaeger-query-bcf5bf78f-ltxjm.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:20 +0000 UTC Normal Pod my-jaeger-query-bcf5bf78f-ltxjm.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:20 +0000 UTC Normal Pod my-jaeger-query-bcf5bf78f-ltxjm.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:20 +0000 UTC Normal Pod my-jaeger-query-bcf5bf78f-ltxjm.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:20 +0000 UTC Normal Pod my-jaeger-query-bcf5bf78f-ltxjm.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:23 +0000 UTC Normal Pod check-span-bh4jz Binding Scheduled Successfully assigned kuttl-test-thorough-reindeer/check-span-bh4jz to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:23 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-bh4jz job-controller logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:23 +0000 UTC Normal Pod report-span-dpdxt Binding Scheduled Successfully assigned kuttl-test-thorough-reindeer/report-span-dpdxt to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:23 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-dpdxt job-controller logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:24 +0000 UTC Normal Pod check-span-bh4jz AddedInterface Add eth0 [10.128.2.32/23] from ovn-kubernetes multus logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:24 +0000 UTC Normal Pod check-span-bh4jz.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:24 +0000 UTC Normal Pod report-span-dpdxt AddedInterface Add eth0 [10.128.2.31/23] from ovn-kubernetes multus logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:24 +0000 UTC Normal Pod report-span-dpdxt.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:25 +0000 UTC Normal Pod report-span-dpdxt.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.237s (1.237s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:25 +0000 UTC Normal Pod report-span-dpdxt.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:25 +0000 UTC Normal Pod report-span-dpdxt.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:26 +0000 UTC Normal Pod check-span-bh4jz.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 2.536s (2.536s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:26 +0000 UTC Normal Pod check-span-bh4jz.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:26 +0000 UTC Normal Pod check-span-bh4jz.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:27 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:27 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:27 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:42 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:42 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-75ccc45495-dhwtk horizontal-pod-autoscaler logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:42 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:09:45 | collector-otlp-production-grpc | 2025-01-20 07:09:45 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:09:45 | collector-otlp-production-grpc | Deleting namespace: kuttl-test-thorough-reindeer === CONT kuttl/harness/collector-autoscale logger.go:42: 07:09:58 | collector-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:09:58 | collector-autoscale | Creating namespace: kuttl-test-pro-gnat logger.go:42: 07:09:58 | collector-autoscale/1-install | starting test step 1-install logger.go:42: 07:09:58 | collector-autoscale/1-install | Jaeger:kuttl-test-pro-gnat/simple-prod created logger.go:42: 07:10:35 | collector-autoscale/1-install | test step completed 1-install logger.go:42: 07:10:35 | collector-autoscale/2- | starting test step 2- logger.go:42: 07:10:35 | collector-autoscale/2- | test step completed 2- logger.go:42: 07:10:35 | collector-autoscale | collector-autoscale events from ns kuttl-test-pro-gnat: logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:05 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprognatsimpleprod-1-bd7c957c-fmt52 Binding Scheduled Successfully assigned kuttl-test-pro-gnat/elasticsearch-cdm-kuttltestprognatsimpleprod-1-bd7c957c-fmt52 to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:05 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestprognatsimpleprod-1-bd7c957c SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestprognatsimpleprod-1-bd7c957c-fmt52 replicaset-controller logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:05 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestprognatsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestprognatsimpleprod-1-bd7c957c to 1 deployment-controller logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:06 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprognatsimpleprod-1-bd7c957c-fmt52 AddedInterface Add eth0 [10.131.0.19/23] from ovn-kubernetes multus logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:06 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprognatsimpleprod-1-bd7c957c-fmt52.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:06 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprognatsimpleprod-1-bd7c957c-fmt52.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:06 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprognatsimpleprod-1-bd7c957c-fmt52.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:06 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprognatsimpleprod-1-bd7c957c-fmt52.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:06 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprognatsimpleprod-1-bd7c957c-fmt52.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:06 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprognatsimpleprod-1-bd7c957c-fmt52.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:21 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestprognatsimpleprod-1-bd7c957c-fmt52.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:32 +0000 UTC Normal Pod simple-prod-collector-cfc999786-8cg46 Binding Scheduled Successfully assigned kuttl-test-pro-gnat/simple-prod-collector-cfc999786-8cg46 to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:32 +0000 UTC Warning Pod simple-prod-collector-cfc999786-8cg46 FailedMount MountVolume.SetUp failed for volume "simple-prod-collector-tls-config-volume" : secret "simple-prod-collector-headless-tls" not found kubelet logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:32 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-cfc999786 SuccessfulCreate Created pod: simple-prod-collector-cfc999786-8cg46 replicaset-controller logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:32 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-cfc999786 to 1 deployment-controller logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:32 +0000 UTC Normal Pod simple-prod-query-b6b995d5b-9rvwk Binding Scheduled Successfully assigned kuttl-test-pro-gnat/simple-prod-query-b6b995d5b-9rvwk to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:32 +0000 UTC Normal ReplicaSet.apps simple-prod-query-b6b995d5b SuccessfulCreate Created pod: simple-prod-query-b6b995d5b-9rvwk replicaset-controller logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:32 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-b6b995d5b to 1 deployment-controller logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:33 +0000 UTC Normal Pod simple-prod-collector-cfc999786-8cg46 AddedInterface Add eth0 [10.129.2.26/23] from ovn-kubernetes multus logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:33 +0000 UTC Normal Pod simple-prod-collector-cfc999786-8cg46.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:33 +0000 UTC Normal Pod simple-prod-collector-cfc999786-8cg46.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:33 +0000 UTC Normal Pod simple-prod-collector-cfc999786-8cg46.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:33 +0000 UTC Normal Pod simple-prod-query-b6b995d5b-9rvwk AddedInterface Add eth0 [10.129.2.27/23] from ovn-kubernetes multus logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:33 +0000 UTC Normal Pod simple-prod-query-b6b995d5b-9rvwk.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:33 +0000 UTC Normal Pod simple-prod-query-b6b995d5b-9rvwk.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:33 +0000 UTC Normal Pod simple-prod-query-b6b995d5b-9rvwk.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:33 +0000 UTC Normal Pod simple-prod-query-b6b995d5b-9rvwk.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:33 +0000 UTC Normal Pod simple-prod-query-b6b995d5b-9rvwk.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:33 +0000 UTC Normal Pod simple-prod-query-b6b995d5b-9rvwk.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:33 +0000 UTC Normal Pod simple-prod-query-b6b995d5b-9rvwk.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:33 +0000 UTC Normal Pod simple-prod-query-b6b995d5b-9rvwk.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:10:35 | collector-autoscale | 2025-01-20 07:10:33 +0000 UTC Normal Pod simple-prod-query-b6b995d5b-9rvwk.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:10:35 | collector-autoscale | Deleting namespace: kuttl-test-pro-gnat === CONT kuttl/harness/collector-otlp-allinone-grpc logger.go:42: 07:10:41 | collector-otlp-allinone-grpc | Creating namespace: kuttl-test-content-kite logger.go:42: 07:10:41 | collector-otlp-allinone-grpc/0-install | starting test step 0-install logger.go:42: 07:10:42 | collector-otlp-allinone-grpc/0-install | Jaeger:kuttl-test-content-kite/my-jaeger created logger.go:42: 07:10:48 | collector-otlp-allinone-grpc/0-install | test step completed 0-install logger.go:42: 07:10:48 | collector-otlp-allinone-grpc/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:10:48 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:10:49 | collector-otlp-allinone-grpc/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:10:57 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:10:58 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:10:58 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/report-span created logger.go:42: 07:10:58 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/check-span created logger.go:42: 07:11:19 | collector-otlp-allinone-grpc/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | collector-otlp-allinone-grpc events from ns kuttl-test-content-kite: logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:46 +0000 UTC Normal Pod my-jaeger-7d76f947ff-lxglz Binding Scheduled Successfully assigned kuttl-test-content-kite/my-jaeger-7d76f947ff-lxglz to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:46 +0000 UTC Normal Pod my-jaeger-7d76f947ff-lxglz AddedInterface Add eth0 [10.131.0.20/23] from ovn-kubernetes multus logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:46 +0000 UTC Normal Pod my-jaeger-7d76f947ff-lxglz.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:46 +0000 UTC Normal Pod my-jaeger-7d76f947ff-lxglz.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:46 +0000 UTC Normal Pod my-jaeger-7d76f947ff-lxglz.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:46 +0000 UTC Normal Pod my-jaeger-7d76f947ff-lxglz.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:46 +0000 UTC Normal Pod my-jaeger-7d76f947ff-lxglz.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:46 +0000 UTC Normal Pod my-jaeger-7d76f947ff-lxglz.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:46 +0000 UTC Normal ReplicaSet.apps my-jaeger-7d76f947ff SuccessfulCreate Created pod: my-jaeger-7d76f947ff-lxglz replicaset-controller logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:46 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-7d76f947ff to 1 deployment-controller logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:54 +0000 UTC Normal Pod my-jaeger-7d76f947ff-lxglz.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:54 +0000 UTC Normal Pod my-jaeger-7d76f947ff-lxglz.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:54 +0000 UTC Normal ReplicaSet.apps my-jaeger-7d76f947ff SuccessfulDelete Deleted pod: my-jaeger-7d76f947ff-lxglz replicaset-controller logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:54 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-7d76f947ff to 0 from 1 deployment-controller logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:55 +0000 UTC Normal Pod my-jaeger-cf899d489-67kbp Binding Scheduled Successfully assigned kuttl-test-content-kite/my-jaeger-cf899d489-67kbp to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:55 +0000 UTC Normal ReplicaSet.apps my-jaeger-cf899d489 SuccessfulCreate Created pod: my-jaeger-cf899d489-67kbp replicaset-controller logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:55 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-cf899d489 to 1 deployment-controller logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:56 +0000 UTC Normal Pod my-jaeger-cf899d489-67kbp AddedInterface Add eth0 [10.131.0.21/23] from ovn-kubernetes multus logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:56 +0000 UTC Normal Pod my-jaeger-cf899d489-67kbp.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:56 +0000 UTC Normal Pod my-jaeger-cf899d489-67kbp.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:56 +0000 UTC Normal Pod my-jaeger-cf899d489-67kbp.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:56 +0000 UTC Normal Pod my-jaeger-cf899d489-67kbp.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:56 +0000 UTC Normal Pod my-jaeger-cf899d489-67kbp.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:56 +0000 UTC Normal Pod my-jaeger-cf899d489-67kbp.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:58 +0000 UTC Normal Pod check-span-nw8nw Binding Scheduled Successfully assigned kuttl-test-content-kite/check-span-nw8nw to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:58 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-nw8nw job-controller logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:58 +0000 UTC Normal Pod report-span-2frnp Binding Scheduled Successfully assigned kuttl-test-content-kite/report-span-2frnp to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:58 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-2frnp job-controller logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:59 +0000 UTC Normal Pod check-span-nw8nw AddedInterface Add eth0 [10.128.2.33/23] from ovn-kubernetes multus logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:59 +0000 UTC Normal Pod check-span-nw8nw.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:59 +0000 UTC Normal Pod check-span-nw8nw.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 745ms (745ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:59 +0000 UTC Normal Pod report-span-2frnp AddedInterface Add eth0 [10.129.2.28/23] from ovn-kubernetes multus logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:59 +0000 UTC Normal Pod report-span-2frnp.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:59 +0000 UTC Normal Pod report-span-2frnp.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 751ms (751ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:59 +0000 UTC Normal Pod report-span-2frnp.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:10:59 +0000 UTC Normal Pod report-span-2frnp.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:11:00 +0000 UTC Normal Pod check-span-nw8nw.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:11:00 +0000 UTC Normal Pod check-span-nw8nw.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | 2025-01-20 07:11:18 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:11:19 | collector-otlp-allinone-grpc | Deleting namespace: kuttl-test-content-kite === CONT kuttl/harness/cassandra-spark logger.go:42: 07:11:32 | cassandra-spark | Ignoring 01-assert.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:11:32 | cassandra-spark | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:11:32 | cassandra-spark | Creating namespace: kuttl-test-bright-sunbeam logger.go:42: 07:11:32 | cassandra-spark | cassandra-spark events from ns kuttl-test-bright-sunbeam: logger.go:42: 07:11:32 | cassandra-spark | Deleting namespace: kuttl-test-bright-sunbeam === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (387.09s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.11s) --- PASS: kuttl/harness/collector-otlp-allinone-http (56.89s) --- PASS: kuttl/harness/set-custom-img (75.77s) --- PASS: kuttl/harness/collector-otlp-production-http (67.86s) --- PASS: kuttl/harness/collector-otlp-production-grpc (80.29s) --- PASS: kuttl/harness/collector-autoscale (43.58s) --- PASS: kuttl/harness/collector-otlp-allinone-grpc (50.32s) --- PASS: kuttl/harness/cassandra-spark (6.12s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name miscellaneous --report --output /logs/artifacts/miscellaneous.xml ./artifacts/kuttl-report.xml time="2025-01-20T07:11:39Z" level=debug msg="Setting a new name for the test suites" time="2025-01-20T07:11:39Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-01-20T07:11:39Z" level=debug msg="normalizing test case names" time="2025-01-20T07:11:39Z" level=debug msg="miscellaneous/artifacts -> miscellaneous_artifacts" time="2025-01-20T07:11:39Z" level=debug msg="miscellaneous/collector-otlp-allinone-http -> miscellaneous_collector_otlp_allinone_http" time="2025-01-20T07:11:39Z" level=debug msg="miscellaneous/set-custom-img -> miscellaneous_set_custom_img" time="2025-01-20T07:11:39Z" level=debug msg="miscellaneous/collector-otlp-production-http -> miscellaneous_collector_otlp_production_http" time="2025-01-20T07:11:39Z" level=debug msg="miscellaneous/collector-otlp-production-grpc -> miscellaneous_collector_otlp_production_grpc" time="2025-01-20T07:11:39Z" level=debug msg="miscellaneous/collector-autoscale -> miscellaneous_collector_autoscale" time="2025-01-20T07:11:39Z" level=debug msg="miscellaneous/collector-otlp-allinone-grpc -> miscellaneous_collector_otlp_allinone_grpc" time="2025-01-20T07:11:39Z" level=debug msg="miscellaneous/cassandra-spark -> miscellaneous_cassandra_spark" +----------------------------------------------+--------+ | NAME | RESULT | +----------------------------------------------+--------+ | miscellaneous_artifacts | passed | | miscellaneous_collector_otlp_allinone_http | passed | | miscellaneous_set_custom_img | passed | | miscellaneous_collector_otlp_production_http | passed | | miscellaneous_collector_otlp_production_grpc | passed | | miscellaneous_collector_autoscale | passed | | miscellaneous_collector_otlp_allinone_grpc | passed | | miscellaneous_cassandra_spark | passed | +----------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh upgrade false true + '[' 3 -ne 3 ']' + test_suite_name=upgrade + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/upgrade.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-upgrade make[2]: Entering directory '/tmp/jaeger-tests' make docker JAEGER_VERSION=1.62.1 IMG="quay.io//jaeger-operator:next" make[3]: Entering directory '/tmp/jaeger-tests' [ ! -z "true" ] || docker build --build-arg=GOPROXY= --build-arg=VERSION="1.62.0" --build-arg=JAEGER_VERSION=1.62.1 --build-arg=TARGETARCH= --build-arg VERSION_DATE=2025-01-20T07:11:39Z --build-arg VERSION_PKG="github.com/jaegertracing/jaeger-operator/pkg/version" -t "quay.io//jaeger-operator:next" . make[3]: Leaving directory '/tmp/jaeger-tests' touch build-e2e-upgrade-image SKIP_ES_EXTERNAL=true IMG=quay.io//jaeger-operator:"1.62.0" JAEGER_OPERATOR_VERSION="1.62.0" JAEGER_VERSION="1.62.0" ./tests/e2e/upgrade/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-15-030107 True False 17m Cluster version is 4.18.0-0.nightly-2025-01-15-030107' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-15-030107 True False 17m Cluster version is 4.18.0-0.nightly-2025-01-15-030107' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/upgrade/render.sh ++ export SUITE_DIR=./tests/e2e/upgrade ++ SUITE_DIR=./tests/e2e/upgrade ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/upgrade ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + export JAEGER_NAME + '[' true = true ']' + skip_test upgrade 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade + warning 'upgrade: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade: Test not supported in OpenShift\e[0m' WAR: upgrade: Test not supported in OpenShift + '[' true = true ']' + skip_test upgrade-from-latest-release 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade-from-latest-release + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade-from-latest-release + warning 'upgrade-from-latest-release: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade-from-latest-release: Test not supported in OpenShift\e[0m' WAR: upgrade-from-latest-release: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running upgrade E2E tests' Running upgrade E2E tests + cd tests/e2e/upgrade/_build + set +e + KUBECONFIG=/tmp/kubeconfig-3197734706 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 07:11:41 | artifacts | Creating namespace: kuttl-test-certain-snipe logger.go:42: 07:11:41 | artifacts | artifacts events from ns kuttl-test-certain-snipe: logger.go:42: 07:11:41 | artifacts | Deleting namespace: kuttl-test-certain-snipe === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (6.36s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.20s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name upgrade --report --output /logs/artifacts/upgrade.xml ./artifacts/kuttl-report.xml time="2025-01-20T07:11:47Z" level=debug msg="Setting a new name for the test suites" time="2025-01-20T07:11:47Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-01-20T07:11:47Z" level=debug msg="normalizing test case names" time="2025-01-20T07:11:47Z" level=debug msg="upgrade/artifacts -> upgrade_artifacts" +-------------------+--------+ | NAME | RESULT | +-------------------+--------+ | upgrade_artifacts | passed | +-------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh elasticsearch false true + '[' 3 -ne 3 ']' + test_suite_name=elasticsearch + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/elasticsearch.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-elasticsearch make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true \ KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ ./tests/e2e/elasticsearch/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-15-030107 True False 17m Cluster version is 4.18.0-0.nightly-2025-01-15-030107' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-15-030107 True False 17m Cluster version is 4.18.0-0.nightly-2025-01-15-030107' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/elasticsearch/render.sh ++ export SUITE_DIR=./tests/e2e/elasticsearch ++ SUITE_DIR=./tests/e2e/elasticsearch ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/elasticsearch ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + start_test es-from-aio-to-production + '[' 1 -ne 1 ']' + test_name=es-from-aio-to-production + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-from-aio-to-production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-from-aio-to-production\e[0m' Rendering files for test es-from-aio-to-production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-from-aio-to-production + cd es-from-aio-to-production + jaeger_name=my-jaeger + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 03 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=03 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./03-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch.redundancyPolicy="ZeroRedundancy"' ./03-install.yaml + render_smoke_test my-jaeger true 04 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=04 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./04-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./04-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test es-increasing-replicas + '[' 1 -ne 1 ']' + test_name=es-increasing-replicas + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-increasing-replicas' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-increasing-replicas\e[0m' Rendering files for test es-increasing-replicas + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-from-aio-to-production + '[' es-from-aio-to-production '!=' _build ']' + cd .. + mkdir -p es-increasing-replicas + cd es-increasing-replicas + jaeger_name=simple-prod + '[' true = true ']' + jaeger_deployment_mode=production_autoprovisioned + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.replicas=2 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.query.replicas=2 ./02-install.yaml + cp ./01-assert.yaml ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.replicas=2 ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .status.readyReplicas=2 ./02-assert.yaml + render_smoke_test simple-prod true 03 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=03 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./03-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + cp ./02-install.yaml ./04-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.elasticsearch.nodeCount=2 ./04-install.yaml + /tmp/jaeger-tests/bin/gomplate -f ./openshift-check-es-nodes.yaml.template -o ./05-check-es-nodes.yaml + '[' true = true ']' + skip_test es-index-cleaner-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-increasing-replicas + '[' es-increasing-replicas '!=' _build ']' + cd .. + rm -rf es-index-cleaner-upstream + warning 'es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_index_cleaner -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-index-cleaner-autoprov + '[' 1 -ne 1 ']' + test_name=es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-index-cleaner-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-index-cleaner-autoprov\e[0m' Rendering files for test es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-index-cleaner-autoprov + cd es-index-cleaner-autoprov + jaeger_name=test-es-index-cleaner-with-prefix + cronjob_name=test-es-index-cleaner-with-prefix-es-index-cleaner + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + cp ../../es-index-cleaner-upstream/04-assert.yaml ../../es-index-cleaner-upstream/README.md . + render_install_jaeger test-es-index-cleaner-with-prefix production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options.es.index-prefix=""' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.enabled=false ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.numberOfDays=0 ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.esIndexCleaner.schedule="*/1 * * * *"' ./01-install.yaml + render_report_spans test-es-index-cleaner-with-prefix true 5 00 true 02 + '[' 6 -ne 6 ']' + jaeger=test-es-index-cleaner-with-prefix + is_secured=true + number_of_spans=5 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + export JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=5 + DAYS=5 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + sed 's~enabled: false~enabled: true~gi' ./01-install.yaml + CRONJOB_NAME=test-es-index-cleaner-with-prefix-es-index-cleaner + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./04-wait-es-index-cleaner.yaml + /tmp/jaeger-tests/bin/gomplate -f ./01-install.yaml -o ./05-install.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 00 06 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=00 + test_step=06 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=test-es-index-cleaner-with-prefix-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=test-es-index-cleaner-with-prefix-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./06-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./06-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.16"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.16"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.16"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.16 ++ version_ge 5.8.16 5.4 +++ echo 5.8.16 5.4 +++ tr ' ' '\n' +++ sort -rV +++ head -n 1 ++ test 5.8.16 == 5.8.16 + '[' -n '' ']' + skip_test es-index-cleaner-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-index-cleaner-autoprov + '[' es-index-cleaner-autoprov '!=' _build ']' + cd .. + rm -rf es-index-cleaner-managed + warning 'es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + start_test es-multiinstance + '[' 1 -ne 1 ']' + test_name=es-multiinstance + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-multiinstance' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-multiinstance\e[0m' Rendering files for test es-multiinstance + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-multiinstance + cd es-multiinstance + jaeger_name=instance-1 + render_install_jaeger instance-1 production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=instance-1 + JAEGER_NAME=instance-1 + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f ./03-create-second-instance.yaml.template -o 03-create-second-instance.yaml + '[' true = true ']' + skip_test es-rollover-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-rollover-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-multiinstance + '[' es-multiinstance '!=' _build ']' + cd .. + rm -rf es-rollover-upstream + warning 'es-rollover-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_rollover -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-rollover-autoprov + '[' 1 -ne 1 ']' + test_name=es-rollover-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-rollover-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-rollover-autoprov\e[0m' Rendering files for test es-rollover-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-rollover-autoprov + cd es-rollover-autoprov + cp ../../es-rollover-upstream/05-assert.yaml ../../es-rollover-upstream/05-install.yaml ../../es-rollover-upstream/README.md . + jaeger_name=my-jaeger + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_report_spans my-jaeger true 2 00 true 02 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 00 03 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=00 + test_step=03 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./03-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./03-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 01 04 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=01 + test_step=04 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=01 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./04-check-indices.yaml + JOB_NUMBER=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./04-assert.yaml + render_report_spans my-jaeger true 2 02 true 06 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=02 + ensure_reported_spans=true + test_step=06 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=02 + JOB_NUMBER=02 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./06-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./06-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 02 07 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=02 + test_step=07 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=02 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./07-check-indices.yaml + JOB_NUMBER=02 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./07-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' 03 08 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + job_number=03 + test_step=08 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=03 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./08-check-indices.yaml + JOB_NUMBER=03 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./08-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' 04 09 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + job_number=04 + test_step=09 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=04 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./09-check-indices.yaml + JOB_NUMBER=04 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./09-assert.yaml + render_report_spans my-jaeger true 2 03 true 10 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=03 + ensure_reported_spans=true + test_step=10 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=03 + JOB_NUMBER=03 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./10-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./10-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + CRONJOB_NAME=my-jaeger-es-rollover + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./11-wait-rollover.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-000002'\'',' 05 11 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-000002'\'',' + job_number=05 + test_step=11 + escape_command ''\''--name'\'', '\''jaeger-span-000002'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-000002'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-000002'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-000002'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=05 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./11-check-indices.yaml + JOB_NUMBER=05 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./11-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' 06 12 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + job_number=06 + test_step=12 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=06 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./12-check-indices.yaml + JOB_NUMBER=06 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./12-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.16"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.16"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.16"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.16 ++ version_ge 5.8.16 5.4 +++ echo 5.8.16 5.4 +++ tr ' ' '\n' +++ sort -rV +++ head -n 1 ++ test 5.8.16 == 5.8.16 + '[' -n '' ']' + skip_test es-rollover-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-rollover-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-rollover-autoprov + '[' es-rollover-autoprov '!=' _build ']' + cd .. + rm -rf es-rollover-managed + warning 'es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + skip_test es-spark-dependencies 'This test is not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=es-spark-dependencies + message='This test is not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + rm -rf es-spark-dependencies + warning 'es-spark-dependencies: This test is not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-spark-dependencies: This test is not supported in OpenShift\e[0m' WAR: es-spark-dependencies: This test is not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running elasticsearch E2E tests' Running elasticsearch E2E tests + cd tests/e2e/elasticsearch/_build + set +e + KUBECONFIG=/tmp/kubeconfig-3197734706 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 7 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/es-from-aio-to-production === PAUSE kuttl/harness/es-from-aio-to-production === RUN kuttl/harness/es-increasing-replicas === PAUSE kuttl/harness/es-increasing-replicas === RUN kuttl/harness/es-index-cleaner-autoprov === PAUSE kuttl/harness/es-index-cleaner-autoprov === RUN kuttl/harness/es-multiinstance === PAUSE kuttl/harness/es-multiinstance === RUN kuttl/harness/es-rollover-autoprov === PAUSE kuttl/harness/es-rollover-autoprov === RUN kuttl/harness/es-simple-prod === PAUSE kuttl/harness/es-simple-prod === CONT kuttl/harness/artifacts logger.go:42: 07:12:11 | artifacts | Creating namespace: kuttl-test-still-bream logger.go:42: 07:12:11 | artifacts | artifacts events from ns kuttl-test-still-bream: logger.go:42: 07:12:11 | artifacts | Deleting namespace: kuttl-test-still-bream === CONT kuttl/harness/es-multiinstance logger.go:42: 07:12:18 | es-multiinstance | Ignoring 03-create-second-instance.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:12:18 | es-multiinstance | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:12:18 | es-multiinstance | Creating namespace: kuttl-test-full-monster logger.go:42: 07:12:18 | es-multiinstance/0-clear-namespace | starting test step 0-clear-namespace logger.go:42: 07:12:18 | es-multiinstance/0-clear-namespace | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --ignore-not-found=true] logger.go:42: 07:12:18 | es-multiinstance/0-clear-namespace | test step completed 0-clear-namespace logger.go:42: 07:12:18 | es-multiinstance/1-install | starting test step 1-install logger.go:42: 07:12:18 | es-multiinstance/1-install | Jaeger:kuttl-test-full-monster/instance-1 created logger.go:42: 07:12:54 | es-multiinstance/1-install | test step completed 1-install logger.go:42: 07:12:54 | es-multiinstance/2-create-namespace | starting test step 2-create-namespace logger.go:42: 07:12:54 | es-multiinstance/2-create-namespace | running command: [sh -c kubectl create namespace jaeger-e2e-multiinstance-test] logger.go:42: 07:12:54 | es-multiinstance/2-create-namespace | namespace/jaeger-e2e-multiinstance-test created logger.go:42: 07:12:54 | es-multiinstance/2-create-namespace | test step completed 2-create-namespace logger.go:42: 07:12:54 | es-multiinstance/3-create-second-instance | starting test step 3-create-second-instance logger.go:42: 07:12:54 | es-multiinstance/3-create-second-instance | running command: [sh -c kubectl apply -f ./01-install.yaml -n jaeger-e2e-multiinstance-test] logger.go:42: 07:12:55 | es-multiinstance/3-create-second-instance | jaeger.jaegertracing.io/instance-1 created logger.go:42: 07:12:55 | es-multiinstance/3-create-second-instance | running command: [sh -c /tmp/jaeger-tests/bin/kubectl-kuttl assert ./01-assert.yaml -n jaeger-e2e-multiinstance-test --timeout 1000] logger.go:42: 07:13:45 | es-multiinstance/3-create-second-instance | assert is valid logger.go:42: 07:13:45 | es-multiinstance/3-create-second-instance | test step completed 3-create-second-instance logger.go:42: 07:13:45 | es-multiinstance/4-check-secrets | starting test step 4-check-secrets logger.go:42: 07:13:45 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n $NAMESPACE > secret1] logger.go:42: 07:13:45 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n jaeger-e2e-multiinstance-test > secret2] logger.go:42: 07:13:45 | es-multiinstance/4-check-secrets | running command: [sh -c cmp --silent secret1 secret2 || exit 0] logger.go:42: 07:13:45 | es-multiinstance/4-check-secrets | test step completed 4-check-secrets logger.go:42: 07:13:45 | es-multiinstance/5-delete | starting test step 5-delete logger.go:42: 07:13:45 | es-multiinstance/5-delete | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --wait=false] logger.go:42: 07:13:45 | es-multiinstance/5-delete | namespace "jaeger-e2e-multiinstance-test" deleted logger.go:42: 07:13:45 | es-multiinstance/5-delete | test step completed 5-delete logger.go:42: 07:13:45 | es-multiinstance | es-multiinstance events from ns kuttl-test-full-monster: logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:12:25 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestfullmonsterinstance1-1-54b6c8cb77 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestfullmonsterinstance1-1-54b6c8cbfkz96 replicaset-controller logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:12:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfullmonsterinstance1-1-54b6c8cbfkz96 Binding Scheduled Successfully assigned kuttl-test-full-monster/elasticsearch-cdm-kuttltestfullmonsterinstance1-1-54b6c8cbfkz96 to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:12:25 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestfullmonsterinstance1-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestfullmonsterinstance1-1-54b6c8cb77 to 1 deployment-controller logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:12:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfullmonsterinstance1-1-54b6c8cbfkz96 AddedInterface Add eth0 [10.131.0.22/23] from ovn-kubernetes multus logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:12:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfullmonsterinstance1-1-54b6c8cbfkz96.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:12:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfullmonsterinstance1-1-54b6c8cbfkz96.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:12:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfullmonsterinstance1-1-54b6c8cbfkz96.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:12:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfullmonsterinstance1-1-54b6c8cbfkz96.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:12:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfullmonsterinstance1-1-54b6c8cbfkz96.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:12:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfullmonsterinstance1-1-54b6c8cbfkz96.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:12:40 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestfullmonsterinstance1-1-54b6c8cbfkz96.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:12:52 +0000 UTC Normal Pod instance-1-collector-77d8946fd6-k5hvp Binding Scheduled Successfully assigned kuttl-test-full-monster/instance-1-collector-77d8946fd6-k5hvp to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:12:52 +0000 UTC Normal Pod instance-1-collector-77d8946fd6-k5hvp AddedInterface Add eth0 [10.128.2.34/23] from ovn-kubernetes multus logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:12:52 +0000 UTC Normal Pod instance-1-collector-77d8946fd6-k5hvp.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:12:52 +0000 UTC Normal Pod instance-1-collector-77d8946fd6-k5hvp.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:12:52 +0000 UTC Normal Pod instance-1-collector-77d8946fd6-k5hvp.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:12:52 +0000 UTC Normal ReplicaSet.apps instance-1-collector-77d8946fd6 SuccessfulCreate Created pod: instance-1-collector-77d8946fd6-k5hvp replicaset-controller logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:12:52 +0000 UTC Normal Deployment.apps instance-1-collector ScalingReplicaSet Scaled up replica set instance-1-collector-77d8946fd6 to 1 deployment-controller logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:12:52 +0000 UTC Normal Pod instance-1-query-5d74b9fd45-rqkmm Binding Scheduled Successfully assigned kuttl-test-full-monster/instance-1-query-5d74b9fd45-rqkmm to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:12:52 +0000 UTC Normal Pod instance-1-query-5d74b9fd45-rqkmm AddedInterface Add eth0 [10.129.2.29/23] from ovn-kubernetes multus logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:12:52 +0000 UTC Normal Pod instance-1-query-5d74b9fd45-rqkmm.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:12:52 +0000 UTC Normal Pod instance-1-query-5d74b9fd45-rqkmm.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:12:52 +0000 UTC Normal Pod instance-1-query-5d74b9fd45-rqkmm.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:12:52 +0000 UTC Normal Pod instance-1-query-5d74b9fd45-rqkmm.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:12:52 +0000 UTC Normal ReplicaSet.apps instance-1-query-5d74b9fd45 SuccessfulCreate Created pod: instance-1-query-5d74b9fd45-rqkmm replicaset-controller logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:12:52 +0000 UTC Normal Deployment.apps instance-1-query ScalingReplicaSet Scaled up replica set instance-1-query-5d74b9fd45 to 1 deployment-controller logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:12:53 +0000 UTC Normal Pod instance-1-query-5d74b9fd45-rqkmm.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:12:53 +0000 UTC Normal Pod instance-1-query-5d74b9fd45-rqkmm.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:12:53 +0000 UTC Normal Pod instance-1-query-5d74b9fd45-rqkmm.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:12:53 +0000 UTC Normal Pod instance-1-query-5d74b9fd45-rqkmm.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:12:53 +0000 UTC Normal Pod instance-1-query-5d74b9fd45-rqkmm.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:13:07 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:13:07 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:13:07 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:13:45 | es-multiinstance | 2025-01-20 07:13:37 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod instance-1-collector-77d8946fd6-k5hvp horizontal-pod-autoscaler logger.go:42: 07:13:46 | es-multiinstance | 2025-01-20 07:13:37 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod instance-1-collector-77d8946fd6-k5hvp horizontal-pod-autoscaler logger.go:42: 07:13:46 | es-multiinstance | 2025-01-20 07:13:37 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod instance-1-collector-77d8946fd6-k5hvp horizontal-pod-autoscaler logger.go:42: 07:13:46 | es-multiinstance | Deleting namespace: kuttl-test-full-monster === CONT kuttl/harness/es-simple-prod logger.go:42: 07:13:53 | es-simple-prod | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:13:53 | es-simple-prod | Creating namespace: kuttl-test-fair-minnow logger.go:42: 07:13:53 | es-simple-prod | es-simple-prod events from ns kuttl-test-fair-minnow: logger.go:42: 07:13:53 | es-simple-prod | Deleting namespace: kuttl-test-fair-minnow === CONT kuttl/harness/es-rollover-autoprov logger.go:42: 07:14:00 | es-rollover-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:14:00 | es-rollover-autoprov | Creating namespace: kuttl-test-one-tahr logger.go:42: 07:14:00 | es-rollover-autoprov/1-install | starting test step 1-install logger.go:42: 07:14:00 | es-rollover-autoprov/1-install | Jaeger:kuttl-test-one-tahr/my-jaeger created logger.go:42: 07:14:37 | es-rollover-autoprov/1-install | test step completed 1-install logger.go:42: 07:14:37 | es-rollover-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 07:14:37 | es-rollover-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:14:39 | es-rollover-autoprov/2-report-spans | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:14:45 | es-rollover-autoprov/2-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 07:14:45 | es-rollover-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 07:14:46 | es-rollover-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 07:15:12 | es-rollover-autoprov/2-report-spans | test step completed 2-report-spans logger.go:42: 07:15:12 | es-rollover-autoprov/3-check-indices | starting test step 3-check-indices logger.go:42: 07:15:12 | es-rollover-autoprov/3-check-indices | Job:kuttl-test-one-tahr/00-check-indices created logger.go:42: 07:15:17 | es-rollover-autoprov/3-check-indices | test step completed 3-check-indices logger.go:42: 07:15:17 | es-rollover-autoprov/4-check-indices | starting test step 4-check-indices logger.go:42: 07:15:17 | es-rollover-autoprov/4-check-indices | Job:kuttl-test-one-tahr/01-check-indices created logger.go:42: 07:15:23 | es-rollover-autoprov/4-check-indices | test step completed 4-check-indices logger.go:42: 07:15:23 | es-rollover-autoprov/5-install | starting test step 5-install logger.go:42: 07:15:24 | es-rollover-autoprov/5-install | Jaeger:kuttl-test-one-tahr/my-jaeger updated logger.go:42: 07:15:42 | es-rollover-autoprov/5-install | test step completed 5-install logger.go:42: 07:15:42 | es-rollover-autoprov/6-report-spans | starting test step 6-report-spans logger.go:42: 07:15:42 | es-rollover-autoprov/6-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:15:50 | es-rollover-autoprov/6-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=02 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-02-job.yaml] logger.go:42: 07:15:51 | es-rollover-autoprov/6-report-spans | running command: [sh -c kubectl apply -f report-span-02-job.yaml -n $NAMESPACE] logger.go:42: 07:15:51 | es-rollover-autoprov/6-report-spans | job.batch/02-report-span created logger.go:42: 07:16:18 | es-rollover-autoprov/6-report-spans | test step completed 6-report-spans logger.go:42: 07:16:18 | es-rollover-autoprov/7-check-indices | starting test step 7-check-indices logger.go:42: 07:16:18 | es-rollover-autoprov/7-check-indices | Job:kuttl-test-one-tahr/02-check-indices created logger.go:42: 07:16:24 | es-rollover-autoprov/7-check-indices | test step completed 7-check-indices logger.go:42: 07:16:24 | es-rollover-autoprov/8-check-indices | starting test step 8-check-indices logger.go:42: 07:16:25 | es-rollover-autoprov/8-check-indices | Job:kuttl-test-one-tahr/03-check-indices created logger.go:42: 07:16:29 | es-rollover-autoprov/8-check-indices | test step completed 8-check-indices logger.go:42: 07:16:29 | es-rollover-autoprov/9-check-indices | starting test step 9-check-indices logger.go:42: 07:16:29 | es-rollover-autoprov/9-check-indices | Job:kuttl-test-one-tahr/04-check-indices created logger.go:42: 07:16:34 | es-rollover-autoprov/9-check-indices | test step completed 9-check-indices logger.go:42: 07:16:34 | es-rollover-autoprov/10-report-spans | starting test step 10-report-spans logger.go:42: 07:16:34 | es-rollover-autoprov/10-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:16:41 | es-rollover-autoprov/10-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=03 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-03-job.yaml] logger.go:42: 07:16:42 | es-rollover-autoprov/10-report-spans | running command: [sh -c kubectl apply -f report-span-03-job.yaml -n $NAMESPACE] logger.go:42: 07:16:42 | es-rollover-autoprov/10-report-spans | job.batch/03-report-span created logger.go:42: 07:17:07 | es-rollover-autoprov/10-report-spans | test step completed 10-report-spans logger.go:42: 07:17:07 | es-rollover-autoprov/11-check-indices | starting test step 11-check-indices logger.go:42: 07:17:07 | es-rollover-autoprov/11-check-indices | running command: [sh -c go run ../../../../cmd-utils/wait-cronjob/main.go --cronjob my-jaeger-es-rollover --namespace $NAMESPACE] logger.go:42: 07:17:17 | es-rollover-autoprov/11-check-indices | time="2025-01-20T07:17:17Z" level=debug msg="Checking if the my-jaeger-es-rollover CronJob exists" logger.go:42: 07:17:17 | es-rollover-autoprov/11-check-indices | time="2025-01-20T07:17:17Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 07:17:17 | es-rollover-autoprov/11-check-indices | time="2025-01-20T07:17:17Z" level=info msg="Cronjob my-jaeger-es-rollover found successfully" logger.go:42: 07:17:17 | es-rollover-autoprov/11-check-indices | time="2025-01-20T07:17:17Z" level=debug msg="Waiting for the next scheduled job from my-jaeger-es-rollover cronjob" logger.go:42: 07:17:17 | es-rollover-autoprov/11-check-indices | time="2025-01-20T07:17:17Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:17:27 | es-rollover-autoprov/11-check-indices | time="2025-01-20T07:17:27Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:17:37 | es-rollover-autoprov/11-check-indices | time="2025-01-20T07:17:37Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:17:47 | es-rollover-autoprov/11-check-indices | time="2025-01-20T07:17:47Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:17:57 | es-rollover-autoprov/11-check-indices | time="2025-01-20T07:17:57Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:18:07 | es-rollover-autoprov/11-check-indices | time="2025-01-20T07:18:07Z" level=info msg="Job of owner my-jaeger-es-rollover succeeded after my-jaeger-es-rollover 50.051094602s" logger.go:42: 07:18:07 | es-rollover-autoprov/11-check-indices | Job:kuttl-test-one-tahr/05-check-indices created logger.go:42: 07:18:16 | es-rollover-autoprov/11-check-indices | test step completed 11-check-indices logger.go:42: 07:18:16 | es-rollover-autoprov/12-check-indices | starting test step 12-check-indices logger.go:42: 07:18:17 | es-rollover-autoprov/12-check-indices | Job:kuttl-test-one-tahr/06-check-indices created logger.go:42: 07:18:22 | es-rollover-autoprov/12-check-indices | test step completed 12-check-indices logger.go:42: 07:18:22 | es-rollover-autoprov | es-rollover-autoprov events from ns kuttl-test-one-tahr: logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:07 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestonetahrmyjaeger-1-5dd8ddc565-pml6p Binding Scheduled Successfully assigned kuttl-test-one-tahr/elasticsearch-cdm-kuttltestonetahrmyjaeger-1-5dd8ddc565-pml6p to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:07 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestonetahrmyjaeger-1-5dd8ddc565-pml6p FailedMount MountVolume.SetUp failed for volume "elasticsearch-metrics" : secret "elasticsearch-metrics" not found kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:07 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestonetahrmyjaeger-1-5dd8ddc565 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestonetahrmyjaeger-1-5dd8ddc565-pml6p replicaset-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:07 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestonetahrmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestonetahrmyjaeger-1-5dd8ddc565 to 1 deployment-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:08 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestonetahrmyjaeger-1-5dd8ddc565-pml6p AddedInterface Add eth0 [10.131.0.23/23] from ovn-kubernetes multus logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:08 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestonetahrmyjaeger-1-5dd8ddc565-pml6p.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:08 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestonetahrmyjaeger-1-5dd8ddc565-pml6p.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:08 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestonetahrmyjaeger-1-5dd8ddc565-pml6p.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:08 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestonetahrmyjaeger-1-5dd8ddc565-pml6p.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:08 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestonetahrmyjaeger-1-5dd8ddc565-pml6p.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:08 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestonetahrmyjaeger-1-5dd8ddc565-pml6p.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:18 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestonetahrmyjaeger-1-5dd8ddc565-pml6p.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:23 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestonetahrmyjaeger-1-5dd8ddc565-pml6p.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:34 +0000 UTC Normal Pod my-jaeger-collector-7b7584768b-p4pz5 Binding Scheduled Successfully assigned kuttl-test-one-tahr/my-jaeger-collector-7b7584768b-p4pz5 to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:34 +0000 UTC Warning Pod my-jaeger-collector-7b7584768b-p4pz5 FailedMount MountVolume.SetUp failed for volume "my-jaeger-collector-tls-config-volume" : secret "my-jaeger-collector-headless-tls" not found kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:34 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7b7584768b SuccessfulCreate Created pod: my-jaeger-collector-7b7584768b-p4pz5 replicaset-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:34 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-7b7584768b to 1 deployment-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:34 +0000 UTC Normal Pod my-jaeger-query-689fb87866-vb5fc Binding Scheduled Successfully assigned kuttl-test-one-tahr/my-jaeger-query-689fb87866-vb5fc to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:34 +0000 UTC Warning Pod my-jaeger-query-689fb87866-vb5fc FailedMount MountVolume.SetUp failed for volume "my-jaeger-ui-oauth-proxy-tls" : secret "my-jaeger-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:34 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-689fb87866 SuccessfulCreate Created pod: my-jaeger-query-689fb87866-vb5fc replicaset-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:34 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-689fb87866 to 1 deployment-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:35 +0000 UTC Normal Pod my-jaeger-collector-7b7584768b-p4pz5 AddedInterface Add eth0 [10.129.2.32/23] from ovn-kubernetes multus logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:35 +0000 UTC Normal Pod my-jaeger-collector-7b7584768b-p4pz5.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:35 +0000 UTC Normal Pod my-jaeger-collector-7b7584768b-p4pz5.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:35 +0000 UTC Normal Pod my-jaeger-collector-7b7584768b-p4pz5.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:35 +0000 UTC Normal Pod my-jaeger-query-689fb87866-vb5fc AddedInterface Add eth0 [10.129.2.33/23] from ovn-kubernetes multus logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:35 +0000 UTC Normal Pod my-jaeger-query-689fb87866-vb5fc.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:35 +0000 UTC Normal Pod my-jaeger-query-689fb87866-vb5fc.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:35 +0000 UTC Normal Pod my-jaeger-query-689fb87866-vb5fc.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:35 +0000 UTC Normal Pod my-jaeger-query-689fb87866-vb5fc.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:35 +0000 UTC Normal Pod my-jaeger-query-689fb87866-vb5fc.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:35 +0000 UTC Normal Pod my-jaeger-query-689fb87866-vb5fc.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:35 +0000 UTC Normal Pod my-jaeger-query-689fb87866-vb5fc.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:36 +0000 UTC Normal Pod my-jaeger-query-689fb87866-vb5fc.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:36 +0000 UTC Normal Pod my-jaeger-query-689fb87866-vb5fc.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:40 +0000 UTC Normal Pod my-jaeger-query-689fb87866-vb5fc.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:40 +0000 UTC Normal Pod my-jaeger-query-689fb87866-vb5fc.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:40 +0000 UTC Normal Pod my-jaeger-query-689fb87866-vb5fc.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:40 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-689fb87866 SuccessfulDelete Deleted pod: my-jaeger-query-689fb87866-vb5fc replicaset-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:40 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-689fb87866 to 0 from 1 deployment-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:41 +0000 UTC Normal Pod my-jaeger-query-5545f59664-lvmbp Binding Scheduled Successfully assigned kuttl-test-one-tahr/my-jaeger-query-5545f59664-lvmbp to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:41 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-5545f59664 SuccessfulCreate Created pod: my-jaeger-query-5545f59664-lvmbp replicaset-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:41 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-5545f59664 to 1 deployment-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:42 +0000 UTC Normal Pod my-jaeger-query-5545f59664-lvmbp AddedInterface Add eth0 [10.129.2.34/23] from ovn-kubernetes multus logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:42 +0000 UTC Normal Pod my-jaeger-query-5545f59664-lvmbp.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:42 +0000 UTC Normal Pod my-jaeger-query-5545f59664-lvmbp.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:42 +0000 UTC Normal Pod my-jaeger-query-5545f59664-lvmbp.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:42 +0000 UTC Normal Pod my-jaeger-query-5545f59664-lvmbp.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:42 +0000 UTC Normal Pod my-jaeger-query-5545f59664-lvmbp.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:42 +0000 UTC Normal Pod my-jaeger-query-5545f59664-lvmbp.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:42 +0000 UTC Normal Pod my-jaeger-query-5545f59664-lvmbp.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:42 +0000 UTC Normal Pod my-jaeger-query-5545f59664-lvmbp.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:42 +0000 UTC Normal Pod my-jaeger-query-5545f59664-lvmbp.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:46 +0000 UTC Normal Pod 00-report-span-w55jh Binding Scheduled Successfully assigned kuttl-test-one-tahr/00-report-span-w55jh to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:46 +0000 UTC Normal Pod 00-report-span-w55jh AddedInterface Add eth0 [10.128.2.36/23] from ovn-kubernetes multus logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:46 +0000 UTC Normal Pod 00-report-span-w55jh.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:46 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-w55jh job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:48 +0000 UTC Normal Pod 00-report-span-w55jh.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 2.087s (2.087s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:48 +0000 UTC Normal Pod 00-report-span-w55jh.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:48 +0000 UTC Normal Pod 00-report-span-w55jh.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:49 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:49 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-7b7584768b-p4pz5 horizontal-pod-autoscaler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:14:49 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:11 +0000 UTC Normal Job.batch 00-report-span Completed Job completed job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:12 +0000 UTC Normal Pod 00-check-indices-swzqf Binding Scheduled Successfully assigned kuttl-test-one-tahr/00-check-indices-swzqf to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:12 +0000 UTC Normal Job.batch 00-check-indices SuccessfulCreate Created pod: 00-check-indices-swzqf job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:13 +0000 UTC Normal Pod 00-check-indices-swzqf AddedInterface Add eth0 [10.128.2.38/23] from ovn-kubernetes multus logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:13 +0000 UTC Normal Pod 00-check-indices-swzqf.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:14 +0000 UTC Normal Pod 00-check-indices-swzqf.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 702ms (702ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:14 +0000 UTC Normal Pod 00-check-indices-swzqf.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:14 +0000 UTC Normal Pod 00-check-indices-swzqf.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:16 +0000 UTC Normal Job.batch 00-check-indices Completed Job completed job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:17 +0000 UTC Normal Pod 01-check-indices-2nc9h Binding Scheduled Successfully assigned kuttl-test-one-tahr/01-check-indices-2nc9h to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:17 +0000 UTC Normal Job.batch 01-check-indices SuccessfulCreate Created pod: 01-check-indices-2nc9h job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:18 +0000 UTC Normal Pod 01-check-indices-2nc9h AddedInterface Add eth0 [10.128.2.39/23] from ovn-kubernetes multus logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:18 +0000 UTC Normal Pod 01-check-indices-2nc9h.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:19 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-7b7584768b-p4pz5 horizontal-pod-autoscaler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:19 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-7b7584768b-p4pz5 horizontal-pod-autoscaler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:20 +0000 UTC Normal Pod 01-check-indices-2nc9h.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 2.43s (2.43s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:20 +0000 UTC Normal Pod 01-check-indices-2nc9h.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:20 +0000 UTC Normal Pod 01-check-indices-2nc9h.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:23 +0000 UTC Normal Job.batch 01-check-indices Completed Job completed job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:25 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-bwv4g Binding Scheduled Successfully assigned kuttl-test-one-tahr/my-jaeger-es-rollover-create-mapping-bwv4g to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:25 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping SuccessfulCreate Created pod: my-jaeger-es-rollover-create-mapping-bwv4g job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:26 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-bwv4g AddedInterface Add eth0 [10.128.2.40/23] from ovn-kubernetes multus logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:26 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-bwv4g.spec.containers{my-jaeger-es-rollover-create-mapping} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:44283d1da2e4839c9a23cf2848ddc67d853a6d1fe532e04c642a6ecee08b25cb" kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:35 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-bwv4g.spec.containers{my-jaeger-es-rollover-create-mapping} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:44283d1da2e4839c9a23cf2848ddc67d853a6d1fe532e04c642a6ecee08b25cb" in 9.875s (9.875s including waiting). Image size: 111742913 bytes. kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:36 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-bwv4g.spec.containers{my-jaeger-es-rollover-create-mapping} Created Created container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:36 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-bwv4g.spec.containers{my-jaeger-es-rollover-create-mapping} Started Started container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:38 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping Completed Job completed job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:39 +0000 UTC Normal Pod my-jaeger-collector-7b7584768b-p4pz5.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:39 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7b7584768b SuccessfulDelete Deleted pod: my-jaeger-collector-7b7584768b-p4pz5 replicaset-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:39 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-7b7584768b to 0 from 1 deployment-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:39 +0000 UTC Normal Pod my-jaeger-query-5545f59664-lvmbp.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:39 +0000 UTC Normal Pod my-jaeger-query-5545f59664-lvmbp.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:39 +0000 UTC Normal Pod my-jaeger-query-5545f59664-lvmbp.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:39 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-5545f59664 SuccessfulDelete Deleted pod: my-jaeger-query-5545f59664-lvmbp replicaset-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:39 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-5545f59664 to 0 from 1 deployment-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:40 +0000 UTC Normal Pod my-jaeger-collector-589494cc55-7qw5x Binding Scheduled Successfully assigned kuttl-test-one-tahr/my-jaeger-collector-589494cc55-7qw5x to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:40 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-589494cc55 SuccessfulCreate Created pod: my-jaeger-collector-589494cc55-7qw5x replicaset-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:40 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-589494cc55 to 1 deployment-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:40 +0000 UTC Normal Pod my-jaeger-query-bf84b8cb5-rcdv4 Binding Scheduled Successfully assigned kuttl-test-one-tahr/my-jaeger-query-bf84b8cb5-rcdv4 to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:40 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-bf84b8cb5 SuccessfulCreate Created pod: my-jaeger-query-bf84b8cb5-rcdv4 replicaset-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:40 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-bf84b8cb5 to 1 deployment-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:41 +0000 UTC Normal Pod my-jaeger-collector-589494cc55-7qw5x AddedInterface Add eth0 [10.129.2.35/23] from ovn-kubernetes multus logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:41 +0000 UTC Normal Pod my-jaeger-collector-589494cc55-7qw5x.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:41 +0000 UTC Normal Pod my-jaeger-collector-589494cc55-7qw5x.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:41 +0000 UTC Normal Pod my-jaeger-collector-589494cc55-7qw5x.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:41 +0000 UTC Normal Pod my-jaeger-query-bf84b8cb5-rcdv4 AddedInterface Add eth0 [10.129.2.36/23] from ovn-kubernetes multus logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:41 +0000 UTC Normal Pod my-jaeger-query-bf84b8cb5-rcdv4.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:41 +0000 UTC Normal Pod my-jaeger-query-bf84b8cb5-rcdv4.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:41 +0000 UTC Normal Pod my-jaeger-query-bf84b8cb5-rcdv4.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:41 +0000 UTC Normal Pod my-jaeger-query-bf84b8cb5-rcdv4.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:41 +0000 UTC Normal Pod my-jaeger-query-bf84b8cb5-rcdv4.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:41 +0000 UTC Normal Pod my-jaeger-query-bf84b8cb5-rcdv4.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:41 +0000 UTC Normal Pod my-jaeger-query-bf84b8cb5-rcdv4.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:41 +0000 UTC Normal Pod my-jaeger-query-bf84b8cb5-rcdv4.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:41 +0000 UTC Normal Pod my-jaeger-query-bf84b8cb5-rcdv4.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:49 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:49 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:49 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:51 +0000 UTC Normal Pod 02-report-span-jz9h7 Binding Scheduled Successfully assigned kuttl-test-one-tahr/02-report-span-jz9h7 to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:51 +0000 UTC Normal Pod 02-report-span-jz9h7 AddedInterface Add eth0 [10.128.2.41/23] from ovn-kubernetes multus logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:51 +0000 UTC Normal Pod 02-report-span-jz9h7.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:51 +0000 UTC Normal Job.batch 02-report-span SuccessfulCreate Created pod: 02-report-span-jz9h7 job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:55 +0000 UTC Normal Pod 02-report-span-jz9h7.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 3.192s (3.192s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:55 +0000 UTC Normal Pod 02-report-span-jz9h7.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:15:55 +0000 UTC Normal Pod 02-report-span-jz9h7.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28955956-c5g58 Binding Scheduled Successfully assigned kuttl-test-one-tahr/my-jaeger-es-lookback-28955956-c5g58 to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28955956-c5g58 AddedInterface Add eth0 [10.128.2.43/23] from ovn-kubernetes multus logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28955956-c5g58.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:44283d1da2e4839c9a23cf2848ddc67d853a6d1fe532e04c642a6ecee08b25cb" already present on machine kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28955956-c5g58.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28955956-c5g58.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28955956 SuccessfulCreate Created pod: my-jaeger-es-lookback-28955956-c5g58 job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28955956 cronjob-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28955956-xbb5d Binding Scheduled Successfully assigned kuttl-test-one-tahr/my-jaeger-es-rollover-28955956-xbb5d to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28955956-xbb5d AddedInterface Add eth0 [10.128.2.42/23] from ovn-kubernetes multus logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28955956-xbb5d.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:44283d1da2e4839c9a23cf2848ddc67d853a6d1fe532e04c642a6ecee08b25cb" already present on machine kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28955956-xbb5d.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28955956-xbb5d.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28955956 SuccessfulCreate Created pod: my-jaeger-es-rollover-28955956-xbb5d job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28955956 cronjob-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:04 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28955956 Completed Job completed job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:04 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28955956, condition: Complete cronjob-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:04 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28955956 Completed Job completed job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:04 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28955956, condition: Complete cronjob-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:17 +0000 UTC Normal Job.batch 02-report-span Completed Job completed job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:18 +0000 UTC Normal Pod 02-check-indices-tnzl9 Binding Scheduled Successfully assigned kuttl-test-one-tahr/02-check-indices-tnzl9 to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:18 +0000 UTC Normal Job.batch 02-check-indices SuccessfulCreate Created pod: 02-check-indices-tnzl9 job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:19 +0000 UTC Normal Pod 02-check-indices-tnzl9 AddedInterface Add eth0 [10.128.2.44/23] from ovn-kubernetes multus logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:19 +0000 UTC Normal Pod 02-check-indices-tnzl9.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:19 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-589494cc55-7qw5x horizontal-pod-autoscaler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:19 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-589494cc55-7qw5x horizontal-pod-autoscaler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:19 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-589494cc55-7qw5x horizontal-pod-autoscaler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:20 +0000 UTC Normal Pod 02-check-indices-tnzl9.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.493s (1.493s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:21 +0000 UTC Normal Pod 02-check-indices-tnzl9.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:21 +0000 UTC Normal Pod 02-check-indices-tnzl9.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:24 +0000 UTC Normal Job.batch 02-check-indices Completed Job completed job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:25 +0000 UTC Normal Pod 03-check-indices-rlkcf Binding Scheduled Successfully assigned kuttl-test-one-tahr/03-check-indices-rlkcf to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:25 +0000 UTC Normal Pod 03-check-indices-rlkcf AddedInterface Add eth0 [10.128.2.45/23] from ovn-kubernetes multus logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:25 +0000 UTC Normal Pod 03-check-indices-rlkcf.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:25 +0000 UTC Normal Job.batch 03-check-indices SuccessfulCreate Created pod: 03-check-indices-rlkcf job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:26 +0000 UTC Normal Pod 03-check-indices-rlkcf.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 460ms (460ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:26 +0000 UTC Normal Pod 03-check-indices-rlkcf.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:26 +0000 UTC Normal Pod 03-check-indices-rlkcf.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:29 +0000 UTC Normal Job.batch 03-check-indices Completed Job completed job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:29 +0000 UTC Normal Pod 04-check-indices-qg8v5 Binding Scheduled Successfully assigned kuttl-test-one-tahr/04-check-indices-qg8v5 to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:29 +0000 UTC Normal Pod 04-check-indices-qg8v5 AddedInterface Add eth0 [10.128.2.46/23] from ovn-kubernetes multus logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:29 +0000 UTC Normal Pod 04-check-indices-qg8v5.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:29 +0000 UTC Normal Job.batch 04-check-indices SuccessfulCreate Created pod: 04-check-indices-qg8v5 job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:31 +0000 UTC Normal Pod 04-check-indices-qg8v5.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.319s (1.319s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:31 +0000 UTC Normal Pod 04-check-indices-qg8v5.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:31 +0000 UTC Normal Pod 04-check-indices-qg8v5.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:34 +0000 UTC Normal Job.batch 04-check-indices Completed Job completed job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:42 +0000 UTC Normal Pod 03-report-span-9rf67 Binding Scheduled Successfully assigned kuttl-test-one-tahr/03-report-span-9rf67 to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:42 +0000 UTC Normal Job.batch 03-report-span SuccessfulCreate Created pod: 03-report-span-9rf67 job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:43 +0000 UTC Normal Pod 03-report-span-9rf67 AddedInterface Add eth0 [10.128.2.47/23] from ovn-kubernetes multus logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:43 +0000 UTC Normal Pod 03-report-span-9rf67.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:43 +0000 UTC Normal Pod 03-report-span-9rf67.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 469ms (469ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:43 +0000 UTC Normal Pod 03-report-span-9rf67.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:16:43 +0000 UTC Normal Pod 03-report-span-9rf67.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:17:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28955957-nwmld Binding Scheduled Successfully assigned kuttl-test-one-tahr/my-jaeger-es-lookback-28955957-nwmld to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:17:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28955957-nwmld AddedInterface Add eth0 [10.128.2.48/23] from ovn-kubernetes multus logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:17:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28955957-nwmld.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:44283d1da2e4839c9a23cf2848ddc67d853a6d1fe532e04c642a6ecee08b25cb" already present on machine kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:17:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28955957-nwmld.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:17:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28955957-nwmld.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:17:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28955957 SuccessfulCreate Created pod: my-jaeger-es-lookback-28955957-nwmld job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:17:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28955957 cronjob-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:17:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28955957-mlc2t Binding Scheduled Successfully assigned kuttl-test-one-tahr/my-jaeger-es-rollover-28955957-mlc2t to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:17:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28955957-mlc2t AddedInterface Add eth0 [10.128.2.49/23] from ovn-kubernetes multus logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:17:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28955957-mlc2t.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:44283d1da2e4839c9a23cf2848ddc67d853a6d1fe532e04c642a6ecee08b25cb" already present on machine kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:17:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28955957-mlc2t.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:17:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28955957-mlc2t.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:17:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28955957 SuccessfulCreate Created pod: my-jaeger-es-rollover-28955957-mlc2t job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:17:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28955957 cronjob-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:17:03 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28955957 Completed Job completed job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:17:03 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28955957, condition: Complete cronjob-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:17:03 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28955957 Completed Job completed job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:17:03 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28955957, condition: Complete cronjob-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:17:06 +0000 UTC Normal Job.batch 03-report-span Completed Job completed job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28955958-h7n7b Binding Scheduled Successfully assigned kuttl-test-one-tahr/my-jaeger-es-lookback-28955958-h7n7b to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28955958-h7n7b AddedInterface Add eth0 [10.128.2.51/23] from ovn-kubernetes multus logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28955958-h7n7b.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:44283d1da2e4839c9a23cf2848ddc67d853a6d1fe532e04c642a6ecee08b25cb" already present on machine kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28955958-h7n7b.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28955958-h7n7b.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28955958 SuccessfulCreate Created pod: my-jaeger-es-lookback-28955958-h7n7b job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28955958 cronjob-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28955958-czxbl Binding Scheduled Successfully assigned kuttl-test-one-tahr/my-jaeger-es-rollover-28955958-czxbl to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28955958-czxbl AddedInterface Add eth0 [10.128.2.50/23] from ovn-kubernetes multus logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28955958-czxbl.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:44283d1da2e4839c9a23cf2848ddc67d853a6d1fe532e04c642a6ecee08b25cb" already present on machine kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28955958-czxbl.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28955958-czxbl.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28955958 SuccessfulCreate Created pod: my-jaeger-es-rollover-28955958-czxbl job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28955958 cronjob-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:03 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28955958 Completed Job completed job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:03 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28955958, condition: Complete cronjob-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:03 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28955958 Completed Job completed job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:03 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28955958, condition: Complete cronjob-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:07 +0000 UTC Normal Pod 05-check-indices-jzxcp Binding Scheduled Successfully assigned kuttl-test-one-tahr/05-check-indices-jzxcp to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:07 +0000 UTC Normal Job.batch 05-check-indices SuccessfulCreate Created pod: 05-check-indices-jzxcp job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:08 +0000 UTC Normal Pod 05-check-indices-jzxcp AddedInterface Add eth0 [10.128.2.52/23] from ovn-kubernetes multus logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:08 +0000 UTC Normal Pod 05-check-indices-jzxcp.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:14 +0000 UTC Normal Pod 05-check-indices-jzxcp.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 5.422s (5.422s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:14 +0000 UTC Normal Pod 05-check-indices-jzxcp.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:14 +0000 UTC Normal Pod 05-check-indices-jzxcp.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:16 +0000 UTC Normal Job.batch 05-check-indices Completed Job completed job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:17 +0000 UTC Normal Pod 06-check-indices-26nth Binding Scheduled Successfully assigned kuttl-test-one-tahr/06-check-indices-26nth to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:17 +0000 UTC Normal Pod 06-check-indices-26nth AddedInterface Add eth0 [10.128.2.53/23] from ovn-kubernetes multus logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:17 +0000 UTC Normal Pod 06-check-indices-26nth.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:17 +0000 UTC Normal Job.batch 06-check-indices SuccessfulCreate Created pod: 06-check-indices-26nth job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:18 +0000 UTC Normal Pod 06-check-indices-26nth.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 718ms (718ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:18 +0000 UTC Normal Pod 06-check-indices-26nth.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:18 +0000 UTC Normal Pod 06-check-indices-26nth.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:18:22 | es-rollover-autoprov | 2025-01-20 07:18:21 +0000 UTC Normal Job.batch 06-check-indices Completed Job completed job-controller logger.go:42: 07:18:22 | es-rollover-autoprov | Deleting namespace: kuttl-test-one-tahr === CONT kuttl/harness/es-increasing-replicas logger.go:42: 07:18:29 | es-increasing-replicas | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:18:29 | es-increasing-replicas | Ignoring check-es-nodes.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:18:29 | es-increasing-replicas | Ignoring openshift-check-es-nodes.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:18:29 | es-increasing-replicas | Creating namespace: kuttl-test-civil-jawfish logger.go:42: 07:18:29 | es-increasing-replicas/1-install | starting test step 1-install logger.go:42: 07:18:29 | es-increasing-replicas/1-install | Jaeger:kuttl-test-civil-jawfish/simple-prod created logger.go:42: 07:19:06 | es-increasing-replicas/1-install | test step completed 1-install logger.go:42: 07:19:06 | es-increasing-replicas/2-install | starting test step 2-install logger.go:42: 07:19:06 | es-increasing-replicas/2-install | Jaeger:kuttl-test-civil-jawfish/simple-prod updated logger.go:42: 07:19:18 | es-increasing-replicas/2-install | test step completed 2-install logger.go:42: 07:19:18 | es-increasing-replicas/3-smoke-test | starting test step 3-smoke-test logger.go:42: 07:19:18 | es-increasing-replicas/3-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:19:19 | es-increasing-replicas/3-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:19:26 | es-increasing-replicas/3-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:19:26 | es-increasing-replicas/3-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:19:27 | es-increasing-replicas/3-smoke-test | job.batch/report-span created logger.go:42: 07:19:27 | es-increasing-replicas/3-smoke-test | job.batch/check-span created logger.go:42: 07:19:40 | es-increasing-replicas/3-smoke-test | test step completed 3-smoke-test logger.go:42: 07:19:40 | es-increasing-replicas/4-install | starting test step 4-install logger.go:42: 07:19:40 | es-increasing-replicas/4-install | Jaeger:kuttl-test-civil-jawfish/simple-prod updated logger.go:42: 07:19:40 | es-increasing-replicas/4-install | test step completed 4-install logger.go:42: 07:19:40 | es-increasing-replicas/5-check-es-nodes | starting test step 5-check-es-nodes logger.go:42: 07:19:40 | es-increasing-replicas/5-check-es-nodes | running command: [sh -c ./check-es-nodes.sh $NAMESPACE] logger.go:42: 07:19:40 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 07:19:40 | es-increasing-replicas/5-check-es-nodes | false logger.go:42: 07:19:40 | es-increasing-replicas/5-check-es-nodes | Error: no matches found logger.go:42: 07:19:45 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 07:19:45 | es-increasing-replicas/5-check-es-nodes | true logger.go:42: 07:19:45 | es-increasing-replicas/5-check-es-nodes | test step completed 5-check-es-nodes logger.go:42: 07:19:45 | es-increasing-replicas | es-increasing-replicas events from ns kuttl-test-civil-jawfish: logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:18:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestciviljawfishsimpleprod-1-55b98822lsv Binding Scheduled Successfully assigned kuttl-test-civil-jawfish/elasticsearch-cdm-kuttltestciviljawfishsimpleprod-1-55b98822lsv to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:18:36 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestciviljawfishsimpleprod-1-55b988c5b5 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestciviljawfishsimpleprod-1-55b98822lsv replicaset-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:18:36 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestciviljawfishsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestciviljawfishsimpleprod-1-55b988c5b5 to 1 deployment-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:18:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestciviljawfishsimpleprod-1-55b98822lsv AddedInterface Add eth0 [10.131.0.24/23] from ovn-kubernetes multus logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:18:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestciviljawfishsimpleprod-1-55b98822lsv.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:18:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestciviljawfishsimpleprod-1-55b98822lsv.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:18:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestciviljawfishsimpleprod-1-55b98822lsv.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:18:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestciviljawfishsimpleprod-1-55b98822lsv.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:18:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestciviljawfishsimpleprod-1-55b98822lsv.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:18:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestciviljawfishsimpleprod-1-55b98822lsv.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:18:52 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestciviljawfishsimpleprod-1-55b98822lsv.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:03 +0000 UTC Normal Pod simple-prod-collector-849f64f7fb-gw77m Binding Scheduled Successfully assigned kuttl-test-civil-jawfish/simple-prod-collector-849f64f7fb-gw77m to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:03 +0000 UTC Normal Pod simple-prod-collector-849f64f7fb-gw77m AddedInterface Add eth0 [10.129.2.37/23] from ovn-kubernetes multus logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:03 +0000 UTC Normal Pod simple-prod-collector-849f64f7fb-gw77m.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:03 +0000 UTC Normal Pod simple-prod-collector-849f64f7fb-gw77m.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:03 +0000 UTC Normal Pod simple-prod-collector-849f64f7fb-gw77m.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:03 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-849f64f7fb SuccessfulCreate Created pod: simple-prod-collector-849f64f7fb-gw77m replicaset-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:03 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-849f64f7fb to 1 deployment-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:03 +0000 UTC Normal Pod simple-prod-query-9d7877d9f-hw5v8 Binding Scheduled Successfully assigned kuttl-test-civil-jawfish/simple-prod-query-9d7877d9f-hw5v8 to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:03 +0000 UTC Normal Pod simple-prod-query-9d7877d9f-hw5v8 AddedInterface Add eth0 [10.129.2.38/23] from ovn-kubernetes multus logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:03 +0000 UTC Normal Pod simple-prod-query-9d7877d9f-hw5v8.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:03 +0000 UTC Normal Pod simple-prod-query-9d7877d9f-hw5v8.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:03 +0000 UTC Normal Pod simple-prod-query-9d7877d9f-hw5v8.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:03 +0000 UTC Normal Pod simple-prod-query-9d7877d9f-hw5v8.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:03 +0000 UTC Normal Pod simple-prod-query-9d7877d9f-hw5v8.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:03 +0000 UTC Normal Pod simple-prod-query-9d7877d9f-hw5v8.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:03 +0000 UTC Normal Pod simple-prod-query-9d7877d9f-hw5v8.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:03 +0000 UTC Normal ReplicaSet.apps simple-prod-query-9d7877d9f SuccessfulCreate Created pod: simple-prod-query-9d7877d9f-hw5v8 replicaset-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:03 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-9d7877d9f to 1 deployment-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:04 +0000 UTC Normal Pod simple-prod-query-9d7877d9f-hw5v8.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:04 +0000 UTC Normal Pod simple-prod-query-9d7877d9f-hw5v8.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:07 +0000 UTC Normal Pod simple-prod-collector-849f64f7fb-xqd48 Binding Scheduled Successfully assigned kuttl-test-civil-jawfish/simple-prod-collector-849f64f7fb-xqd48 to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:07 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-849f64f7fb SuccessfulCreate Created pod: simple-prod-collector-849f64f7fb-xqd48 replicaset-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:07 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-849f64f7fb to 2 from 1 deployment-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:07 +0000 UTC Normal Pod simple-prod-query-9d7877d9f-x8dc6 Binding Scheduled Successfully assigned kuttl-test-civil-jawfish/simple-prod-query-9d7877d9f-x8dc6 to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:07 +0000 UTC Normal ReplicaSet.apps simple-prod-query-9d7877d9f SuccessfulCreate Created pod: simple-prod-query-9d7877d9f-x8dc6 replicaset-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:07 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-9d7877d9f to 2 from 1 deployment-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:08 +0000 UTC Normal Pod simple-prod-collector-849f64f7fb-xqd48 AddedInterface Add eth0 [10.128.2.54/23] from ovn-kubernetes multus logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:08 +0000 UTC Normal Pod simple-prod-collector-849f64f7fb-xqd48.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:08 +0000 UTC Normal Pod simple-prod-collector-849f64f7fb-xqd48.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:08 +0000 UTC Normal Pod simple-prod-collector-849f64f7fb-xqd48.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:08 +0000 UTC Normal Pod simple-prod-query-9d7877d9f-x8dc6 AddedInterface Add eth0 [10.128.2.55/23] from ovn-kubernetes multus logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:08 +0000 UTC Normal Pod simple-prod-query-9d7877d9f-x8dc6.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:11 +0000 UTC Normal Pod simple-prod-query-9d7877d9f-x8dc6.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" in 2.729s (2.729s including waiting). Image size: 192936692 bytes. kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:11 +0000 UTC Normal Pod simple-prod-query-9d7877d9f-x8dc6.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:11 +0000 UTC Normal Pod simple-prod-query-9d7877d9f-x8dc6.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:11 +0000 UTC Normal Pod simple-prod-query-9d7877d9f-x8dc6.spec.containers{oauth-proxy} Pulling Pulling image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:14 +0000 UTC Normal Pod simple-prod-query-9d7877d9f-x8dc6.spec.containers{oauth-proxy} Pulled Successfully pulled image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" in 3.738s (3.738s including waiting). Image size: 339954870 bytes. kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:14 +0000 UTC Normal Pod simple-prod-query-9d7877d9f-x8dc6.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:14 +0000 UTC Normal Pod simple-prod-query-9d7877d9f-x8dc6.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:14 +0000 UTC Normal Pod simple-prod-query-9d7877d9f-x8dc6.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:17 +0000 UTC Normal Pod simple-prod-query-9d7877d9f-x8dc6.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" in 2.301s (2.301s including waiting). Image size: 115697219 bytes. kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:17 +0000 UTC Normal Pod simple-prod-query-9d7877d9f-x8dc6.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:17 +0000 UTC Normal Pod simple-prod-query-9d7877d9f-x8dc6.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:21 +0000 UTC Normal Pod simple-prod-query-9d7877d9f-hw5v8.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:21 +0000 UTC Normal Pod simple-prod-query-9d7877d9f-hw5v8.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:21 +0000 UTC Normal Pod simple-prod-query-9d7877d9f-hw5v8.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:21 +0000 UTC Normal Pod simple-prod-query-9d7877d9f-x8dc6.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:21 +0000 UTC Normal Pod simple-prod-query-9d7877d9f-x8dc6.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:21 +0000 UTC Normal Pod simple-prod-query-9d7877d9f-x8dc6.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:21 +0000 UTC Normal ReplicaSet.apps simple-prod-query-9d7877d9f SuccessfulDelete Deleted pod: simple-prod-query-9d7877d9f-hw5v8 replicaset-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:21 +0000 UTC Normal ReplicaSet.apps simple-prod-query-9d7877d9f SuccessfulDelete Deleted pod: simple-prod-query-9d7877d9f-x8dc6 replicaset-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:21 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-9d7877d9f to 0 from 2 deployment-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:22 +0000 UTC Normal Pod simple-prod-query-64dc7dd749-j88cf Binding Scheduled Successfully assigned kuttl-test-civil-jawfish/simple-prod-query-64dc7dd749-j88cf to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:22 +0000 UTC Normal Pod simple-prod-query-64dc7dd749-rt8qw Binding Scheduled Successfully assigned kuttl-test-civil-jawfish/simple-prod-query-64dc7dd749-rt8qw to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:22 +0000 UTC Normal ReplicaSet.apps simple-prod-query-64dc7dd749 SuccessfulCreate Created pod: simple-prod-query-64dc7dd749-rt8qw replicaset-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:22 +0000 UTC Normal ReplicaSet.apps simple-prod-query-64dc7dd749 SuccessfulCreate Created pod: simple-prod-query-64dc7dd749-j88cf replicaset-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:22 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-64dc7dd749 to 2 deployment-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:23 +0000 UTC Normal Pod simple-prod-query-64dc7dd749-j88cf AddedInterface Add eth0 [10.128.2.56/23] from ovn-kubernetes multus logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:23 +0000 UTC Normal Pod simple-prod-query-64dc7dd749-j88cf.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:23 +0000 UTC Normal Pod simple-prod-query-64dc7dd749-j88cf.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:23 +0000 UTC Normal Pod simple-prod-query-64dc7dd749-j88cf.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:23 +0000 UTC Normal Pod simple-prod-query-64dc7dd749-j88cf.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:23 +0000 UTC Normal Pod simple-prod-query-64dc7dd749-j88cf.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:23 +0000 UTC Normal Pod simple-prod-query-64dc7dd749-j88cf.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:23 +0000 UTC Normal Pod simple-prod-query-64dc7dd749-j88cf.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:23 +0000 UTC Normal Pod simple-prod-query-64dc7dd749-j88cf.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:23 +0000 UTC Normal Pod simple-prod-query-64dc7dd749-j88cf.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:23 +0000 UTC Normal Pod simple-prod-query-64dc7dd749-rt8qw AddedInterface Add eth0 [10.129.2.39/23] from ovn-kubernetes multus logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:23 +0000 UTC Normal Pod simple-prod-query-64dc7dd749-rt8qw.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:23 +0000 UTC Normal Pod simple-prod-query-64dc7dd749-rt8qw.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:23 +0000 UTC Normal Pod simple-prod-query-64dc7dd749-rt8qw.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:23 +0000 UTC Normal Pod simple-prod-query-64dc7dd749-rt8qw.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:23 +0000 UTC Normal Pod simple-prod-query-64dc7dd749-rt8qw.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:23 +0000 UTC Normal Pod simple-prod-query-64dc7dd749-rt8qw.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:23 +0000 UTC Normal Pod simple-prod-query-64dc7dd749-rt8qw.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:23 +0000 UTC Normal Pod simple-prod-query-64dc7dd749-rt8qw.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:23 +0000 UTC Normal Pod simple-prod-query-64dc7dd749-rt8qw.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:27 +0000 UTC Normal Pod check-span-kt9cj Binding Scheduled Successfully assigned kuttl-test-civil-jawfish/check-span-kt9cj to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:27 +0000 UTC Normal Pod check-span-kt9cj AddedInterface Add eth0 [10.128.2.57/23] from ovn-kubernetes multus logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:27 +0000 UTC Normal Pod check-span-kt9cj.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:27 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-kt9cj job-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:27 +0000 UTC Normal Pod report-span-g5qz8 Binding Scheduled Successfully assigned kuttl-test-civil-jawfish/report-span-g5qz8 to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:27 +0000 UTC Normal Pod report-span-g5qz8 AddedInterface Add eth0 [10.129.2.40/23] from ovn-kubernetes multus logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:27 +0000 UTC Normal Pod report-span-g5qz8.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:27 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-g5qz8 job-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:28 +0000 UTC Normal Pod check-span-kt9cj.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.039s (1.039s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:28 +0000 UTC Normal Pod check-span-kt9cj.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:28 +0000 UTC Normal Pod check-span-kt9cj.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:28 +0000 UTC Normal Pod report-span-g5qz8.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 971ms (971ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:28 +0000 UTC Normal Pod report-span-g5qz8.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:28 +0000 UTC Normal Pod report-span-g5qz8.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:39 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:41 +0000 UTC Normal Pod simple-prod-collector-849f64f7fb-gw77m.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:41 +0000 UTC Normal Pod simple-prod-collector-849f64f7fb-xqd48.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:41 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-849f64f7fb SuccessfulDelete Deleted pod: simple-prod-collector-849f64f7fb-xqd48 replicaset-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:41 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled down replica set simple-prod-collector-849f64f7fb to 0 from 2 deployment-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:41 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-64dc7dd749 to 0 from 2 deployment-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:42 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestciviljawfishsimpleprod-2-dc7b748c9 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestciviljawfishsimpleprod-2-dc7b74xq6dd replicaset-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestciviljawfishsimpleprod-2-dc7b74xq6dd Binding Scheduled Successfully assigned kuttl-test-civil-jawfish/elasticsearch-cdm-kuttltestciviljawfishsimpleprod-2-dc7b74xq6dd to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestciviljawfishsimpleprod-2-dc7b74xq6dd AddedInterface Add eth0 [10.128.2.58/23] from ovn-kubernetes multus logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestciviljawfishsimpleprod-2-dc7b74xq6dd.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestciviljawfishsimpleprod-2-dc7b74xq6dd.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestciviljawfishsimpleprod-2-dc7b74xq6dd.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestciviljawfishsimpleprod-2-dc7b74xq6dd.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestciviljawfishsimpleprod-2-dc7b74xq6dd.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:42 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestciviljawfishsimpleprod-2-dc7b74xq6dd.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:42 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestciviljawfishsimpleprod-2 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestciviljawfishsimpleprod-2-dc7b748c9 to 1 deployment-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:42 +0000 UTC Normal Pod simple-prod-collector-79f575995-plczq Binding Scheduled Successfully assigned kuttl-test-civil-jawfish/simple-prod-collector-79f575995-plczq to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:42 +0000 UTC Normal Pod simple-prod-collector-79f575995-xnpdw Binding Scheduled Successfully assigned kuttl-test-civil-jawfish/simple-prod-collector-79f575995-xnpdw to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:42 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-79f575995 SuccessfulCreate Created pod: simple-prod-collector-79f575995-xnpdw replicaset-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:42 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-79f575995 SuccessfulCreate Created pod: simple-prod-collector-79f575995-plczq replicaset-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:42 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-849f64f7fb SuccessfulDelete Deleted pod: simple-prod-collector-849f64f7fb-gw77m replicaset-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:42 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-79f575995 to 2 deployment-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:42 +0000 UTC Normal Pod simple-prod-query-64dc7dd749-j88cf.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:42 +0000 UTC Normal Pod simple-prod-query-64dc7dd749-j88cf.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:42 +0000 UTC Normal Pod simple-prod-query-64dc7dd749-j88cf.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:42 +0000 UTC Normal Pod simple-prod-query-64dc7dd749-rt8qw.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:42 +0000 UTC Normal Pod simple-prod-query-64dc7dd749-rt8qw.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:42 +0000 UTC Normal Pod simple-prod-query-64dc7dd749-rt8qw.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:42 +0000 UTC Normal ReplicaSet.apps simple-prod-query-64dc7dd749 SuccessfulDelete Deleted pod: simple-prod-query-64dc7dd749-rt8qw replicaset-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:42 +0000 UTC Normal ReplicaSet.apps simple-prod-query-64dc7dd749 SuccessfulDelete Deleted pod: simple-prod-query-64dc7dd749-j88cf replicaset-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:42 +0000 UTC Normal Pod simple-prod-query-6ddf5d5bc4-9ptk4 Binding Scheduled Successfully assigned kuttl-test-civil-jawfish/simple-prod-query-6ddf5d5bc4-9ptk4 to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:42 +0000 UTC Normal Pod simple-prod-query-6ddf5d5bc4-b77j5 Binding Scheduled Successfully assigned kuttl-test-civil-jawfish/simple-prod-query-6ddf5d5bc4-b77j5 to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:42 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6ddf5d5bc4 SuccessfulCreate Created pod: simple-prod-query-6ddf5d5bc4-b77j5 replicaset-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:42 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6ddf5d5bc4 SuccessfulCreate Created pod: simple-prod-query-6ddf5d5bc4-9ptk4 replicaset-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:42 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-6ddf5d5bc4 to 2 deployment-controller logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:43 +0000 UTC Normal Pod simple-prod-collector-79f575995-plczq AddedInterface Add eth0 [10.128.2.59/23] from ovn-kubernetes multus logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:43 +0000 UTC Normal Pod simple-prod-collector-79f575995-plczq.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:43 +0000 UTC Normal Pod simple-prod-collector-79f575995-plczq.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:43 +0000 UTC Normal Pod simple-prod-collector-79f575995-plczq.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:43 +0000 UTC Normal Pod simple-prod-collector-79f575995-xnpdw AddedInterface Add eth0 [10.129.2.41/23] from ovn-kubernetes multus logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:43 +0000 UTC Normal Pod simple-prod-collector-79f575995-xnpdw.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:43 +0000 UTC Normal Pod simple-prod-collector-79f575995-xnpdw.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:43 +0000 UTC Normal Pod simple-prod-collector-79f575995-xnpdw.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:43 +0000 UTC Normal Pod simple-prod-query-6ddf5d5bc4-9ptk4 AddedInterface Add eth0 [10.128.2.60/23] from ovn-kubernetes multus logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:43 +0000 UTC Normal Pod simple-prod-query-6ddf5d5bc4-9ptk4.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:43 +0000 UTC Normal Pod simple-prod-query-6ddf5d5bc4-9ptk4.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:43 +0000 UTC Normal Pod simple-prod-query-6ddf5d5bc4-9ptk4.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:43 +0000 UTC Normal Pod simple-prod-query-6ddf5d5bc4-9ptk4.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:43 +0000 UTC Normal Pod simple-prod-query-6ddf5d5bc4-9ptk4.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:43 +0000 UTC Normal Pod simple-prod-query-6ddf5d5bc4-9ptk4.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:43 +0000 UTC Normal Pod simple-prod-query-6ddf5d5bc4-9ptk4.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:43 +0000 UTC Normal Pod simple-prod-query-6ddf5d5bc4-b77j5 AddedInterface Add eth0 [10.129.2.42/23] from ovn-kubernetes multus logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:43 +0000 UTC Normal Pod simple-prod-query-6ddf5d5bc4-b77j5.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:43 +0000 UTC Normal Pod simple-prod-query-6ddf5d5bc4-b77j5.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:43 +0000 UTC Normal Pod simple-prod-query-6ddf5d5bc4-b77j5.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:43 +0000 UTC Normal Pod simple-prod-query-6ddf5d5bc4-b77j5.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:43 +0000 UTC Normal Pod simple-prod-query-6ddf5d5bc4-b77j5.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:43 +0000 UTC Normal Pod simple-prod-query-6ddf5d5bc4-b77j5.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:43 +0000 UTC Normal Pod simple-prod-query-6ddf5d5bc4-b77j5.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:43 +0000 UTC Normal Pod simple-prod-query-6ddf5d5bc4-b77j5.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:43 +0000 UTC Normal Pod simple-prod-query-6ddf5d5bc4-b77j5.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:44 +0000 UTC Normal Pod simple-prod-query-6ddf5d5bc4-9ptk4.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:19:45 | es-increasing-replicas | 2025-01-20 07:19:44 +0000 UTC Normal Pod simple-prod-query-6ddf5d5bc4-9ptk4.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:19:45 | es-increasing-replicas | Deleting namespace: kuttl-test-civil-jawfish === CONT kuttl/harness/es-index-cleaner-autoprov logger.go:42: 07:20:20 | es-index-cleaner-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:20:20 | es-index-cleaner-autoprov | Creating namespace: kuttl-test-viable-crab logger.go:42: 07:20:20 | es-index-cleaner-autoprov/1-install | starting test step 1-install logger.go:42: 07:20:20 | es-index-cleaner-autoprov/1-install | Jaeger:kuttl-test-viable-crab/test-es-index-cleaner-with-prefix created logger.go:42: 07:20:55 | es-index-cleaner-autoprov/1-install | test step completed 1-install logger.go:42: 07:20:55 | es-index-cleaner-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 07:20:55 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE test-es-index-cleaner-with-prefix /dev/null] logger.go:42: 07:20:57 | es-index-cleaner-autoprov/2-report-spans | Warning: resource jaegers/test-es-index-cleaner-with-prefix is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:21:03 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c DAYS=5 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 07:21:04 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 07:21:04 | es-index-cleaner-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 07:31:04 | es-index-cleaner-autoprov/2-report-spans | test step failed 2-report-spans case.go:364: failed in step 2-report-spans case.go:366: --- Job:kuttl-test-viable-crab/00-report-span +++ Job:kuttl-test-viable-crab/00-report-span @@ -1,8 +1,162 @@ apiVersion: batch/v1 kind: Job metadata: + annotations: + kubectl.kubernetes.io/last-applied-configuration: | + {"apiVersion":"batch/v1","kind":"Job","metadata":{"annotations":{},"name":"00-report-span","namespace":"kuttl-test-viable-crab"},"spec":{"backoffLimit":15,"template":{"spec":{"containers":[{"command":["./reporter","--days","5","--verbose"],"env":[{"name":"JAEGER_SERVICE_NAME","value":"smoke-test-service"},{"name":"OPERATION_NAME","value":"smoke-test-operation"},{"name":"JAEGER_ENDPOINT","value":"http://test-es-index-cleaner-with-prefix-collector-headless:14268/api/traces"},{"name":"JAEGER_QUERY","value":"https://test-es-index-cleaner-with-prefix-query/api/traces"},{"name":"SECRET_PATH","value":"/var/run/secrets/api-token/token"}],"image":"quay.io/rhn_support_ikanse/jaeger-asserts:latest","name":"asserts-container","volumeMounts":[{"mountPath":"/var/run/secrets/api-token","name":"token-api-volume"}]}],"restartPolicy":"OnFailure","volumes":[{"name":"token-api-volume","secret":{"secretName":"e2e-test"}}]}}}} + labels: + batch.kubernetes.io/controller-uid: 14feb631-a6bc-4ff2-afa8-ca6af6ae4ba4 + batch.kubernetes.io/job-name: 00-report-span + controller-uid: 14feb631-a6bc-4ff2-afa8-ca6af6ae4ba4 + job-name: 00-report-span + managedFields: + - apiVersion: batch/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + .: {} + f:kubectl.kubernetes.io/last-applied-configuration: {} + f:spec: + f:backoffLimit: {} + f:completionMode: {} + f:completions: {} + f:manualSelector: {} + f:parallelism: {} + f:podReplacementPolicy: {} + f:suspend: {} + f:template: + f:spec: + f:containers: + k:{"name":"asserts-container"}: + .: {} + f:command: {} + f:env: + .: {} + k:{"name":"JAEGER_ENDPOINT"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"JAEGER_QUERY"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"JAEGER_SERVICE_NAME"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"OPERATION_NAME"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"SECRET_PATH"}: + .: {} + f:name: {} + f:value: {} + f:image: {} + f:imagePullPolicy: {} + f:name: {} + f:resources: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/var/run/secrets/api-token"}: + .: {} + f:mountPath: {} + f:name: {} + f:dnsPolicy: {} + f:restartPolicy: {} + f:schedulerName: {} + f:securityContext: {} + f:terminationGracePeriodSeconds: {} + f:volumes: + .: {} + k:{"name":"token-api-volume"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + manager: kubectl-client-side-apply + operation: Update + time: "2025-01-20T07:21:04Z" + - apiVersion: batch/v1 + fieldsType: FieldsV1 + fieldsV1: + f:status: + f:active: {} + f:ready: {} + f:startTime: {} + f:terminating: {} + f:uncountedTerminatedPods: {} + manager: kube-controller-manager + operation: Update + subresource: status + time: "2025-01-20T07:21:08Z" name: 00-report-span namespace: kuttl-test-viable-crab +spec: + backoffLimit: 15 + completionMode: NonIndexed + completions: 1 + manualSelector: false + parallelism: 1 + podReplacementPolicy: TerminatingOrFailed + selector: + matchLabels: + batch.kubernetes.io/controller-uid: 14feb631-a6bc-4ff2-afa8-ca6af6ae4ba4 + suspend: false + template: + metadata: + creationTimestamp: null + labels: + batch.kubernetes.io/controller-uid: 14feb631-a6bc-4ff2-afa8-ca6af6ae4ba4 + batch.kubernetes.io/job-name: 00-report-span + controller-uid: 14feb631-a6bc-4ff2-afa8-ca6af6ae4ba4 + job-name: 00-report-span + spec: + containers: + - command: + - ./reporter + - --days + - "5" + - --verbose + env: + - name: JAEGER_SERVICE_NAME + value: smoke-test-service + - name: OPERATION_NAME + value: smoke-test-operation + - name: JAEGER_ENDPOINT + value: http://test-es-index-cleaner-with-prefix-collector-headless:14268/api/traces + - name: JAEGER_QUERY + value: https://test-es-index-cleaner-with-prefix-query/api/traces + - name: SECRET_PATH + value: /var/run/secrets/api-token/token + image: quay.io/rhn_support_ikanse/jaeger-asserts:latest + imagePullPolicy: Always + name: asserts-container + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /var/run/secrets/api-token + name: token-api-volume + dnsPolicy: ClusterFirst + restartPolicy: OnFailure + schedulerName: default-scheduler + securityContext: {} + terminationGracePeriodSeconds: 30 + volumes: + - name: token-api-volume + secret: + defaultMode: 420 + secretName: e2e-test status: - succeeded: 1 + active: 1 + ready: 1 + startTime: "2025-01-20T07:21:04Z" + terminating: 0 + uncountedTerminatedPods: {} case.go:366: resource Job:kuttl-test-viable-crab/00-report-span: .status.succeeded: key is missing from map logger.go:42: 07:31:04 | es-index-cleaner-autoprov | es-index-cleaner-autoprov events from ns kuttl-test-viable-crab: logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:25 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestviablecrabtestesindexcleane-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestviablecrabtestesindexcleane-1-6f6d4d9487 to 1 deployment-controller logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:26 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestviablecrabtestesindexcleane-1-6f6d4d9487 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestviablecrabtestesindexcleane-1-6qb98d replicaset-controller logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestviablecrabtestesindexcleane-1-6qb98d Binding Scheduled Successfully assigned kuttl-test-viable-crab/elasticsearch-cdm-kuttltestviablecrabtestesindexcleane-1-6qb98d to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestviablecrabtestesindexcleane-1-6qb98d AddedInterface Add eth0 [10.131.0.25/23] from ovn-kubernetes multus logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestviablecrabtestesindexcleane-1-6qb98d.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestviablecrabtestesindexcleane-1-6qb98d.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestviablecrabtestesindexcleane-1-6qb98d.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestviablecrabtestesindexcleane-1-6qb98d.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestviablecrabtestesindexcleane-1-6qb98d.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestviablecrabtestesindexcleane-1-6qb98d.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:36 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestviablecrabtestesindexcleane-1-6qb98d.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:41 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestviablecrabtestesindexcleane-1-6qb98d.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:52 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-666d4589bb-gkl9h Binding Scheduled Successfully assigned kuttl-test-viable-crab/test-es-index-cleaner-with-prefix-collector-666d4589bb-gkl9h to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:52 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-collector-666d4589bb SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-collector-666d4589bb-gkl9h replicaset-controller logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:52 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-collector ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-collector-666d4589bb to 1 deployment-controller logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:52 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-64f96d7466-wgp92 Binding Scheduled Successfully assigned kuttl-test-viable-crab/test-es-index-cleaner-with-prefix-query-64f96d7466-wgp92 to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:52 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-64f96d7466 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-query-64f96d7466-wgp92 replicaset-controller logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:52 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-query-64f96d7466 to 1 deployment-controller logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-666d4589bb-gkl9h AddedInterface Add eth0 [10.129.2.43/23] from ovn-kubernetes multus logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-666d4589bb-gkl9h.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-666d4589bb-gkl9h.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-666d4589bb-gkl9h.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-64f96d7466-wgp92 AddedInterface Add eth0 [10.128.2.61/23] from ovn-kubernetes multus logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-64f96d7466-wgp92.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-64f96d7466-wgp92.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-64f96d7466-wgp92.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-64f96d7466-wgp92.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-64f96d7466-wgp92.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-64f96d7466-wgp92.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-64f96d7466-wgp92.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-64f96d7466-wgp92.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:20:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-64f96d7466-wgp92.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:21:04 +0000 UTC Normal Pod 00-report-span-sspw8 Binding Scheduled Successfully assigned kuttl-test-viable-crab/00-report-span-sspw8 to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:21:04 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-sspw8 job-controller logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:21:05 +0000 UTC Normal Pod 00-report-span-sspw8 AddedInterface Add eth0 [10.129.2.44/23] from ovn-kubernetes multus logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:21:05 +0000 UTC Normal Pod 00-report-span-sspw8.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:21:07 +0000 UTC Normal Pod 00-report-span-sspw8.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 2.012s (2.012s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:21:07 +0000 UTC Normal Pod 00-report-span-sspw8.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:21:07 +0000 UTC Normal Pod 00-report-span-sspw8.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:21:10 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:21:10 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:21:10 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:21:25 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:21:25 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-666d4589bb-gkl9h horizontal-pod-autoscaler logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:21:25 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:21:40 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-666d4589bb-gkl9h horizontal-pod-autoscaler logger.go:42: 07:31:04 | es-index-cleaner-autoprov | 2025-01-20 07:21:40 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-666d4589bb-gkl9h horizontal-pod-autoscaler logger.go:42: 07:31:04 | es-index-cleaner-autoprov | Deleting namespace: kuttl-test-viable-crab === CONT kuttl/harness/es-from-aio-to-production logger.go:42: 07:31:12 | es-from-aio-to-production | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:31:12 | es-from-aio-to-production | Creating namespace: kuttl-test-improved-grouper logger.go:42: 07:31:12 | es-from-aio-to-production/0-install | starting test step 0-install logger.go:42: 07:31:12 | es-from-aio-to-production/0-install | Jaeger:kuttl-test-improved-grouper/my-jaeger created logger.go:42: 07:31:18 | es-from-aio-to-production/0-install | test step completed 0-install logger.go:42: 07:31:18 | es-from-aio-to-production/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:31:18 | es-from-aio-to-production/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:31:19 | es-from-aio-to-production/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:31:27 | es-from-aio-to-production/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:31:27 | es-from-aio-to-production/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:31:27 | es-from-aio-to-production/1-smoke-test | job.batch/report-span created logger.go:42: 07:31:27 | es-from-aio-to-production/1-smoke-test | job.batch/check-span created logger.go:42: 07:31:41 | es-from-aio-to-production/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:31:41 | es-from-aio-to-production/3-install | starting test step 3-install logger.go:42: 07:31:41 | es-from-aio-to-production/3-install | Jaeger:kuttl-test-improved-grouper/my-jaeger updated logger.go:42: 07:32:15 | es-from-aio-to-production/3-install | test step completed 3-install logger.go:42: 07:32:15 | es-from-aio-to-production/4-smoke-test | starting test step 4-smoke-test logger.go:42: 07:32:15 | es-from-aio-to-production/4-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:32:22 | es-from-aio-to-production/4-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:32:23 | es-from-aio-to-production/4-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:32:23 | es-from-aio-to-production/4-smoke-test | job.batch/report-span unchanged logger.go:42: 07:32:23 | es-from-aio-to-production/4-smoke-test | job.batch/check-span unchanged logger.go:42: 07:32:23 | es-from-aio-to-production/4-smoke-test | test step completed 4-smoke-test logger.go:42: 07:32:23 | es-from-aio-to-production | es-from-aio-to-production events from ns kuttl-test-improved-grouper: logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:16 +0000 UTC Normal Pod my-jaeger-d5748fd74-5d748 Binding Scheduled Successfully assigned kuttl-test-improved-grouper/my-jaeger-d5748fd74-5d748 to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:16 +0000 UTC Normal Pod my-jaeger-d5748fd74-5d748 AddedInterface Add eth0 [10.131.0.26/23] from ovn-kubernetes multus logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:16 +0000 UTC Normal Pod my-jaeger-d5748fd74-5d748.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:16 +0000 UTC Normal Pod my-jaeger-d5748fd74-5d748.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:16 +0000 UTC Normal ReplicaSet.apps my-jaeger-d5748fd74 SuccessfulCreate Created pod: my-jaeger-d5748fd74-5d748 replicaset-controller logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:16 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-d5748fd74 to 1 deployment-controller logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:17 +0000 UTC Normal Pod my-jaeger-d5748fd74-5d748.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:17 +0000 UTC Normal Pod my-jaeger-d5748fd74-5d748.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:17 +0000 UTC Normal Pod my-jaeger-d5748fd74-5d748.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:17 +0000 UTC Normal Pod my-jaeger-d5748fd74-5d748.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:24 +0000 UTC Normal Pod my-jaeger-5bfdc94cf6-zqpg6 Binding Scheduled Successfully assigned kuttl-test-improved-grouper/my-jaeger-5bfdc94cf6-zqpg6 to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:24 +0000 UTC Normal ReplicaSet.apps my-jaeger-5bfdc94cf6 SuccessfulCreate Created pod: my-jaeger-5bfdc94cf6-zqpg6 replicaset-controller logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:24 +0000 UTC Normal Pod my-jaeger-d5748fd74-5d748.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:24 +0000 UTC Normal Pod my-jaeger-d5748fd74-5d748.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:24 +0000 UTC Normal ReplicaSet.apps my-jaeger-d5748fd74 SuccessfulDelete Deleted pod: my-jaeger-d5748fd74-5d748 replicaset-controller logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:24 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-d5748fd74 to 0 from 1 deployment-controller logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:24 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-5bfdc94cf6 to 1 deployment-controller logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:25 +0000 UTC Normal Pod my-jaeger-5bfdc94cf6-zqpg6 AddedInterface Add eth0 [10.131.0.27/23] from ovn-kubernetes multus logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:25 +0000 UTC Normal Pod my-jaeger-5bfdc94cf6-zqpg6.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:25 +0000 UTC Normal Pod my-jaeger-5bfdc94cf6-zqpg6.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:25 +0000 UTC Normal Pod my-jaeger-5bfdc94cf6-zqpg6.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:25 +0000 UTC Normal Pod my-jaeger-5bfdc94cf6-zqpg6.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:25 +0000 UTC Normal Pod my-jaeger-5bfdc94cf6-zqpg6.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:25 +0000 UTC Normal Pod my-jaeger-5bfdc94cf6-zqpg6.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:27 +0000 UTC Normal Pod check-span-ksps5 Binding Scheduled Successfully assigned kuttl-test-improved-grouper/check-span-ksps5 to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:27 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-ksps5 job-controller logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:27 +0000 UTC Normal Pod report-span-bktqv Binding Scheduled Successfully assigned kuttl-test-improved-grouper/report-span-bktqv to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:27 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-bktqv job-controller logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:28 +0000 UTC Normal Pod check-span-ksps5 AddedInterface Add eth0 [10.128.2.63/23] from ovn-kubernetes multus logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:28 +0000 UTC Normal Pod check-span-ksps5.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:28 +0000 UTC Normal Pod report-span-bktqv AddedInterface Add eth0 [10.129.2.45/23] from ovn-kubernetes multus logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:28 +0000 UTC Normal Pod report-span-bktqv.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:29 +0000 UTC Normal Pod check-span-ksps5.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 835ms (835ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:29 +0000 UTC Normal Pod check-span-ksps5.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:29 +0000 UTC Normal Pod check-span-ksps5.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:29 +0000 UTC Normal Pod report-span-bktqv.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.296s (1.296s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:29 +0000 UTC Normal Pod report-span-bktqv.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:29 +0000 UTC Normal Pod report-span-bktqv.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:40 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:45 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestimprovedgroupermyjaeger-1-796784648c SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestimprovedgroupermyjaeger-1-79678pzmph replicaset-controller logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestimprovedgroupermyjaeger-1-79678pzmph Binding Scheduled Successfully assigned kuttl-test-improved-grouper/elasticsearch-cdm-kuttltestimprovedgroupermyjaeger-1-79678pzmph to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestimprovedgroupermyjaeger-1-79678pzmph AddedInterface Add eth0 [10.128.2.64/23] from ovn-kubernetes multus logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestimprovedgroupermyjaeger-1-79678pzmph.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestimprovedgroupermyjaeger-1-79678pzmph.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestimprovedgroupermyjaeger-1-79678pzmph.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestimprovedgroupermyjaeger-1-79678pzmph.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestimprovedgroupermyjaeger-1-79678pzmph.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestimprovedgroupermyjaeger-1-79678pzmph.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:45 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestimprovedgroupermyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestimprovedgroupermyjaeger-1-796784648c to 1 deployment-controller logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:31:55 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestimprovedgroupermyjaeger-1-79678pzmph.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:32:00 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestimprovedgroupermyjaeger-1-79678pzmph.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:32:02 +0000 UTC Normal Job.batch report-span Completed Job completed job-controller logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:32:11 +0000 UTC Normal Pod my-jaeger-collector-679dc4c5d7-jlmdh Binding Scheduled Successfully assigned kuttl-test-improved-grouper/my-jaeger-collector-679dc4c5d7-jlmdh to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:32:11 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-679dc4c5d7 SuccessfulCreate Created pod: my-jaeger-collector-679dc4c5d7-jlmdh replicaset-controller logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:32:11 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-679dc4c5d7 to 1 deployment-controller logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:32:11 +0000 UTC Normal Pod my-jaeger-query-658d4d95dd-pmw29 Binding Scheduled Successfully assigned kuttl-test-improved-grouper/my-jaeger-query-658d4d95dd-pmw29 to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:32:11 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-658d4d95dd SuccessfulCreate Created pod: my-jaeger-query-658d4d95dd-pmw29 replicaset-controller logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:32:11 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-658d4d95dd to 1 deployment-controller logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:32:12 +0000 UTC Normal Pod my-jaeger-5bfdc94cf6-zqpg6.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:32:12 +0000 UTC Normal Pod my-jaeger-5bfdc94cf6-zqpg6.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:32:12 +0000 UTC Normal Pod my-jaeger-collector-679dc4c5d7-jlmdh AddedInterface Add eth0 [10.129.2.46/23] from ovn-kubernetes multus logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:32:12 +0000 UTC Normal Pod my-jaeger-collector-679dc4c5d7-jlmdh.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:32:12 +0000 UTC Normal Pod my-jaeger-query-658d4d95dd-pmw29 AddedInterface Add eth0 [10.129.2.47/23] from ovn-kubernetes multus logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:32:12 +0000 UTC Normal Pod my-jaeger-query-658d4d95dd-pmw29.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:32:13 +0000 UTC Normal Pod my-jaeger-collector-679dc4c5d7-jlmdh.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:32:13 +0000 UTC Normal Pod my-jaeger-collector-679dc4c5d7-jlmdh.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:32:13 +0000 UTC Normal Pod my-jaeger-query-658d4d95dd-pmw29.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:32:13 +0000 UTC Normal Pod my-jaeger-query-658d4d95dd-pmw29.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:32:13 +0000 UTC Normal Pod my-jaeger-query-658d4d95dd-pmw29.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:32:13 +0000 UTC Normal Pod my-jaeger-query-658d4d95dd-pmw29.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:32:13 +0000 UTC Normal Pod my-jaeger-query-658d4d95dd-pmw29.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:32:13 +0000 UTC Normal Pod my-jaeger-query-658d4d95dd-pmw29.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:32:13 +0000 UTC Normal Pod my-jaeger-query-658d4d95dd-pmw29.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | 2025-01-20 07:32:13 +0000 UTC Normal Pod my-jaeger-query-658d4d95dd-pmw29.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:32:23 | es-from-aio-to-production | Deleting namespace: kuttl-test-improved-grouper === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- FAIL: kuttl (1218.43s) --- FAIL: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.62s) --- PASS: kuttl/harness/es-multiinstance (95.46s) --- PASS: kuttl/harness/es-simple-prod (6.69s) --- PASS: kuttl/harness/es-rollover-autoprov (269.15s) --- PASS: kuttl/harness/es-increasing-replicas (111.04s) --- FAIL: kuttl/harness/es-index-cleaner-autoprov (651.36s) --- PASS: kuttl/harness/es-from-aio-to-production (77.96s) FAIL + exit_code=1 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name elasticsearch --report --output /logs/artifacts/elasticsearch.xml ./artifacts/kuttl-report.xml time="2025-01-20T07:32:30Z" level=debug msg="Setting a new name for the test suites" time="2025-01-20T07:32:30Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-01-20T07:32:30Z" level=debug msg="normalizing test case names" time="2025-01-20T07:32:30Z" level=debug msg="elasticsearch/artifacts -> elasticsearch_artifacts" time="2025-01-20T07:32:30Z" level=debug msg="elasticsearch/es-multiinstance -> elasticsearch_es_multiinstance" time="2025-01-20T07:32:30Z" level=debug msg="elasticsearch/es-simple-prod -> elasticsearch_es_simple_prod" time="2025-01-20T07:32:30Z" level=debug msg="elasticsearch/es-rollover-autoprov -> elasticsearch_es_rollover_autoprov" time="2025-01-20T07:32:30Z" level=debug msg="elasticsearch/es-increasing-replicas -> elasticsearch_es_increasing_replicas" time="2025-01-20T07:32:30Z" level=debug msg="elasticsearch/es-index-cleaner-autoprov -> elasticsearch_es_index_cleaner_autoprov" time="2025-01-20T07:32:30Z" level=debug msg="elasticsearch/es-from-aio-to-production -> elasticsearch_es_from_aio_to_production" +-----------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------+--------+ | elasticsearch_artifacts | passed | | elasticsearch_es_multiinstance | passed | | elasticsearch_es_simple_prod | passed | | elasticsearch_es_rollover_autoprov | passed | | elasticsearch_es_increasing_replicas | passed | | elasticsearch_es_index_cleaner_autoprov | failed | | elasticsearch_es_from_aio_to_production | passed | +-----------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh examples false true + '[' 3 -ne 3 ']' + test_suite_name=examples + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/examples.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-examples make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ VERTX_IMG=jaegertracing/vertx-create-span:operator-e2e-tests \ ./tests/e2e/examples/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-15-030107 True False 38m Cluster version is 4.18.0-0.nightly-2025-01-15-030107' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-15-030107 True False 38m Cluster version is 4.18.0-0.nightly-2025-01-15-030107' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/examples/render.sh ++ export SUITE_DIR=./tests/e2e/examples ++ SUITE_DIR=./tests/e2e/examples ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/examples ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test examples-agent-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-agent-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-with-priority-class\e[0m' Rendering files for test examples-agent-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + mkdir -p examples-agent-with-priority-class + cd examples-agent-with-priority-class + example_name=agent-with-priority-class + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + '[' true '!=' true ']' + render_install_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + install_file=./02-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/agent-with-priority-class.yaml -o ./02-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./02-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./02-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./02-install.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + JAEGER_NAME=agent-as-daemonset + local jaeger_strategy ++ get_jaeger_strategy ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./02-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./02-install.yaml ++ strategy=DaemonSet ++ '[' DaemonSet = null ']' ++ echo DaemonSet ++ return 0 + jaeger_strategy=DaemonSet + '[' DaemonSet = DaemonSet ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./02-assert.yaml + render_smoke_test_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + jaeger_name=agent-as-daemonset + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test agent-as-daemonset true 02 + '[' 3 -ne 3 ']' + jaeger=agent-as-daemonset + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + export JAEGER_NAME=agent-as-daemonset + JAEGER_NAME=agent-as-daemonset + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-all-in-one-with-options + '[' 1 -ne 1 ']' + test_name=examples-all-in-one-with-options + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-all-in-one-with-options' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-all-in-one-with-options\e[0m' Rendering files for test examples-all-in-one-with-options + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-with-priority-class + '[' examples-agent-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-all-in-one-with-options + cd examples-all-in-one-with-options + example_name=all-in-one-with-options + render_install_example all-in-one-with-options 00 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/all-in-one-with-options.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + JAEGER_NAME=my-jaeger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.metadata.name="my-jaeger"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i 'del(.spec.allInOne.image)' ./00-install.yaml + render_smoke_test_example all-in-one-with-options 01 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + jaeger_name=my-jaeger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + sed -i s~my-jaeger-query:443~my-jaeger-query:443/jaeger~gi ./01-smoke-test.yaml + start_test examples-business-application-injected-sidecar + '[' 1 -ne 1 ']' + test_name=examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-business-application-injected-sidecar' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-business-application-injected-sidecar\e[0m' Rendering files for test examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-all-in-one-with-options + '[' examples-all-in-one-with-options '!=' _build ']' + cd .. + mkdir -p examples-business-application-injected-sidecar + cd examples-business-application-injected-sidecar + example_name=simplest + cp /tmp/jaeger-tests/examples/business-application-injected-sidecar.yaml ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].image=strenv(VERTX_IMG)' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.path="/"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.port=8080' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.initialDelaySeconds=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.failureThreshold=3' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.periodSeconds=10' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.successThreshold=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.timeoutSeconds=1' ./00-install.yaml + render_install_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example simplest 02 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 02 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-collector-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-collector-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-collector-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-collector-with-priority-class\e[0m' Rendering files for test examples-collector-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-business-application-injected-sidecar + '[' examples-business-application-injected-sidecar '!=' _build ']' + cd .. + mkdir -p examples-collector-with-priority-class + cd examples-collector-with-priority-class + example_name=collector-with-priority-class + render_install_example collector-with-priority-class 00 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/collector-with-priority-class.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + JAEGER_NAME=collector-with-high-priority + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example collector-with-priority-class 01 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + jaeger_name=collector-with-high-priority + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test collector-with-high-priority true 01 + '[' 3 -ne 3 ']' + jaeger=collector-with-high-priority + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + export JAEGER_NAME=collector-with-high-priority + JAEGER_NAME=collector-with-high-priority + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-service-types + '[' 1 -ne 1 ']' + test_name=examples-service-types + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-service-types' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-service-types\e[0m' Rendering files for test examples-service-types + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-collector-with-priority-class + '[' examples-collector-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-service-types + cd examples-service-types + example_name=service-types + render_install_example service-types 00 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/service-types.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + JAEGER_NAME=service-types + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example service-types 01 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/service-types.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/service-types.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/service-types.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/service-types.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + jaeger_name=service-types + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test service-types true 01 + '[' 3 -ne 3 ']' + jaeger=service-types + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + export JAEGER_NAME=service-types + JAEGER_NAME=service-types + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod + '[' 1 -ne 1 ']' + test_name=examples-simple-prod + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod\e[0m' Rendering files for test examples-simple-prod + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-service-types + '[' examples-service-types '!=' _build ']' + cd .. + mkdir -p examples-simple-prod + cd examples-simple-prod + example_name=simple-prod + render_install_example simple-prod 01 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod 02 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod-with-volumes + '[' 1 -ne 1 ']' + test_name=examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod-with-volumes' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod-with-volumes\e[0m' Rendering files for test examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod + '[' examples-simple-prod '!=' _build ']' + cd .. + mkdir -p examples-simple-prod-with-volumes + cd examples-simple-prod-with-volumes + example_name=simple-prod-with-volumes + render_install_example simple-prod-with-volumes 01 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod-with-volumes 02 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + /tmp/jaeger-tests/bin/gomplate -f ./03-check-volume.yaml.template -o 03-check-volume.yaml + start_test examples-simplest + '[' 1 -ne 1 ']' + test_name=examples-simplest + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simplest' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simplest\e[0m' Rendering files for test examples-simplest + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod-with-volumes + '[' examples-simple-prod-with-volumes '!=' _build ']' + cd .. + mkdir -p examples-simplest + cd examples-simplest + example_name=simplest + render_install_example simplest 00 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 01 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger + '[' 1 -ne 1 ']' + test_name=examples-with-badger + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger\e[0m' Rendering files for test examples-with-badger + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simplest + '[' examples-simplest '!=' _build ']' + cd .. + mkdir -p examples-with-badger + cd examples-with-badger + example_name=with-badger + render_install_example with-badger 00 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + JAEGER_NAME=with-badger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger 01 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + jaeger_name=with-badger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + export JAEGER_NAME=with-badger + JAEGER_NAME=with-badger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger-and-volume + '[' 1 -ne 1 ']' + test_name=examples-with-badger-and-volume + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger-and-volume' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger-and-volume\e[0m' Rendering files for test examples-with-badger-and-volume + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger + '[' examples-with-badger '!=' _build ']' + cd .. + mkdir -p examples-with-badger-and-volume + cd examples-with-badger-and-volume + example_name=with-badger-and-volume + render_install_example with-badger-and-volume 00 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger-and-volume.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + JAEGER_NAME=with-badger-and-volume + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger-and-volume 01 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + jaeger_name=with-badger-and-volume + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger-and-volume true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger-and-volume + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + export JAEGER_NAME=with-badger-and-volume + JAEGER_NAME=with-badger-and-volume + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-cassandra + '[' 1 -ne 1 ']' + test_name=examples-with-cassandra + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-cassandra' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-cassandra\e[0m' Rendering files for test examples-with-cassandra + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger-and-volume + '[' examples-with-badger-and-volume '!=' _build ']' + cd .. + mkdir -p examples-with-cassandra + cd examples-with-cassandra + example_name=with-cassandra + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-cassandra 01 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-cassandra.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + JAEGER_NAME=with-cassandra + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-cassandra 02 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-cassandra.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-cassandra.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + jaeger_name=with-cassandra + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-cassandra true 02 + '[' 3 -ne 3 ']' + jaeger=with-cassandra + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + export JAEGER_NAME=with-cassandra + JAEGER_NAME=with-cassandra + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-sampling + '[' 1 -ne 1 ']' + test_name=examples-with-sampling + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-sampling' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-sampling\e[0m' Rendering files for test examples-with-sampling + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-cassandra + '[' examples-with-cassandra '!=' _build ']' + cd .. + mkdir -p examples-with-sampling + cd examples-with-sampling + export example_name=with-sampling + example_name=with-sampling + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-sampling 01 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-sampling.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + JAEGER_NAME=with-sampling + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-sampling 02 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-sampling.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-sampling.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + jaeger_name=with-sampling + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-sampling true 02 + '[' 3 -ne 3 ']' + jaeger=with-sampling + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + export JAEGER_NAME=with-sampling + JAEGER_NAME=with-sampling + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-agent-as-daemonset + '[' 1 -ne 1 ']' + test_name=examples-agent-as-daemonset + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-as-daemonset' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-as-daemonset\e[0m' Rendering files for test examples-agent-as-daemonset + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-sampling + '[' examples-with-sampling '!=' _build ']' + cd .. + mkdir -p examples-agent-as-daemonset + cd examples-agent-as-daemonset + '[' true = true ']' + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/agent-as-daemonset.yaml -o 02-install.yaml + '[' true = true ']' + start_test examples-openshift-with-htpasswd + '[' 1 -ne 1 ']' + test_name=examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-openshift-with-htpasswd' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-openshift-with-htpasswd\e[0m' Rendering files for test examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-as-daemonset + '[' examples-agent-as-daemonset '!=' _build ']' + cd .. + mkdir -p examples-openshift-with-htpasswd + cd examples-openshift-with-htpasswd + export JAEGER_NAME=with-htpasswd + JAEGER_NAME=with-htpasswd + export JAEGER_USERNAME=awesomeuser + JAEGER_USERNAME=awesomeuser + export JAEGER_PASSWORD=awesomepassword + JAEGER_PASSWORD=awesomepassword + export 'JAEGER_USER_PASSWORD_HASH=awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' + JAEGER_USER_PASSWORD_HASH='awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ echo 'awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ base64 + SECRET=YXdlc29tZXVzZXI6e1NIQX11VWRxUFZVeXFOQm1FUlUwUXhqM0tGYVpuanc9Cg== + /tmp/jaeger-tests/bin/gomplate -f ./00-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/with-htpasswd.yaml -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + export 'GET_URL_COMMAND=kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + export 'URL=https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + INSECURE=true + JAEGER_USERNAME= + JAEGER_PASSWORD= + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-unsecured.yaml + JAEGER_USERNAME=wronguser + JAEGER_PASSWORD=wrongpassword + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-check-unauthorized.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./04-check-authorized.yaml + skip_test examples-agent-as-daemonset 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-agent-as-daemonset + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-openshift-with-htpasswd + '[' examples-openshift-with-htpasswd '!=' _build ']' + cd .. + rm -rf examples-agent-as-daemonset + warning 'examples-agent-as-daemonset: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-agent-as-daemonset: This test is flaky in Prow CI\e[0m' WAR: examples-agent-as-daemonset: This test is flaky in Prow CI + skip_test examples-with-badger-and-volume 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-with-badger-and-volume + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + rm -rf examples-with-badger-and-volume + warning 'examples-with-badger-and-volume: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-with-badger-and-volume: This test is flaky in Prow CI\e[0m' WAR: examples-with-badger-and-volume: This test is flaky in Prow CI + skip_test examples-collector-with-priority-class 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-collector-with-priority-class + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + rm -rf examples-collector-with-priority-class + warning 'examples-collector-with-priority-class: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-collector-with-priority-class: This test is flaky in Prow CI\e[0m' WAR: examples-collector-with-priority-class: This test is flaky in Prow CI make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running examples E2E tests' Running examples E2E tests + cd tests/e2e/examples/_build + set +e + KUBECONFIG=/tmp/kubeconfig-3197734706 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 12 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/examples-agent-with-priority-class === PAUSE kuttl/harness/examples-agent-with-priority-class === RUN kuttl/harness/examples-all-in-one-with-options === PAUSE kuttl/harness/examples-all-in-one-with-options === RUN kuttl/harness/examples-business-application-injected-sidecar === PAUSE kuttl/harness/examples-business-application-injected-sidecar === RUN kuttl/harness/examples-openshift-with-htpasswd === PAUSE kuttl/harness/examples-openshift-with-htpasswd === RUN kuttl/harness/examples-service-types === PAUSE kuttl/harness/examples-service-types === RUN kuttl/harness/examples-simple-prod === PAUSE kuttl/harness/examples-simple-prod === RUN kuttl/harness/examples-simple-prod-with-volumes === PAUSE kuttl/harness/examples-simple-prod-with-volumes === RUN kuttl/harness/examples-simplest === PAUSE kuttl/harness/examples-simplest === RUN kuttl/harness/examples-with-badger === PAUSE kuttl/harness/examples-with-badger === RUN kuttl/harness/examples-with-cassandra === PAUSE kuttl/harness/examples-with-cassandra === RUN kuttl/harness/examples-with-sampling === PAUSE kuttl/harness/examples-with-sampling === CONT kuttl/harness/artifacts logger.go:42: 07:33:01 | artifacts | Creating namespace: kuttl-test-quality-moose logger.go:42: 07:33:01 | artifacts | artifacts events from ns kuttl-test-quality-moose: logger.go:42: 07:33:01 | artifacts | Deleting namespace: kuttl-test-quality-moose === CONT kuttl/harness/examples-simple-prod logger.go:42: 07:33:07 | examples-simple-prod | Creating namespace: kuttl-test-relaxing-gator logger.go:42: 07:33:07 | examples-simple-prod/1-install | starting test step 1-install logger.go:42: 07:33:08 | examples-simple-prod/1-install | Jaeger:kuttl-test-relaxing-gator/simple-prod created logger.go:42: 07:33:44 | examples-simple-prod/1-install | test step completed 1-install logger.go:42: 07:33:44 | examples-simple-prod/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:33:44 | examples-simple-prod/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:33:45 | examples-simple-prod/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:33:51 | examples-simple-prod/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:33:52 | examples-simple-prod/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:33:52 | examples-simple-prod/2-smoke-test | job.batch/report-span created logger.go:42: 07:33:52 | examples-simple-prod/2-smoke-test | job.batch/check-span created logger.go:42: 07:34:06 | examples-simple-prod/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:34:06 | examples-simple-prod | examples-simple-prod events from ns kuttl-test-relaxing-gator: logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:13 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestrelaxinggatorsimpleprod-1-56f99f5c4c SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestrelaxinggatorsimpleprod-1-56f99hs9vl replicaset-controller logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:13 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrelaxinggatorsimpleprod-1-56f99hs9vl Binding Scheduled Successfully assigned kuttl-test-relaxing-gator/elasticsearch-cdm-kuttltestrelaxinggatorsimpleprod-1-56f99hs9vl to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:13 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestrelaxinggatorsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestrelaxinggatorsimpleprod-1-56f99f5c4c to 1 deployment-controller logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:14 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrelaxinggatorsimpleprod-1-56f99hs9vl AddedInterface Add eth0 [10.131.0.28/23] from ovn-kubernetes multus logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:14 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrelaxinggatorsimpleprod-1-56f99hs9vl.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:14 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrelaxinggatorsimpleprod-1-56f99hs9vl.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:14 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrelaxinggatorsimpleprod-1-56f99hs9vl.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:14 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrelaxinggatorsimpleprod-1-56f99hs9vl.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:14 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrelaxinggatorsimpleprod-1-56f99hs9vl.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:14 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrelaxinggatorsimpleprod-1-56f99hs9vl.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:24 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestrelaxinggatorsimpleprod-1-56f99hs9vl.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:29 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestrelaxinggatorsimpleprod-1-56f99hs9vl.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:40 +0000 UTC Normal Pod simple-prod-collector-5bb84d889b-rpjrt Binding Scheduled Successfully assigned kuttl-test-relaxing-gator/simple-prod-collector-5bb84d889b-rpjrt to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:40 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-5bb84d889b SuccessfulCreate Created pod: simple-prod-collector-5bb84d889b-rpjrt replicaset-controller logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:40 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-5bb84d889b to 1 deployment-controller logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:40 +0000 UTC Normal Pod simple-prod-query-6569c74cf6-rhtq8 Binding Scheduled Successfully assigned kuttl-test-relaxing-gator/simple-prod-query-6569c74cf6-rhtq8 to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:40 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6569c74cf6 SuccessfulCreate Created pod: simple-prod-query-6569c74cf6-rhtq8 replicaset-controller logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:40 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-6569c74cf6 to 1 deployment-controller logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:41 +0000 UTC Normal Pod simple-prod-collector-5bb84d889b-rpjrt AddedInterface Add eth0 [10.129.2.48/23] from ovn-kubernetes multus logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:41 +0000 UTC Normal Pod simple-prod-collector-5bb84d889b-rpjrt.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:41 +0000 UTC Normal Pod simple-prod-collector-5bb84d889b-rpjrt.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:41 +0000 UTC Normal Pod simple-prod-collector-5bb84d889b-rpjrt.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:41 +0000 UTC Normal Pod simple-prod-query-6569c74cf6-rhtq8 AddedInterface Add eth0 [10.128.2.65/23] from ovn-kubernetes multus logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:41 +0000 UTC Normal Pod simple-prod-query-6569c74cf6-rhtq8.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:41 +0000 UTC Normal Pod simple-prod-query-6569c74cf6-rhtq8.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:41 +0000 UTC Normal Pod simple-prod-query-6569c74cf6-rhtq8.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:41 +0000 UTC Normal Pod simple-prod-query-6569c74cf6-rhtq8.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:41 +0000 UTC Normal Pod simple-prod-query-6569c74cf6-rhtq8.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:41 +0000 UTC Normal Pod simple-prod-query-6569c74cf6-rhtq8.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:41 +0000 UTC Normal Pod simple-prod-query-6569c74cf6-rhtq8.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:41 +0000 UTC Normal Pod simple-prod-query-6569c74cf6-rhtq8.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:41 +0000 UTC Normal Pod simple-prod-query-6569c74cf6-rhtq8.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:47 +0000 UTC Normal Pod simple-prod-query-6569c74cf6-rhtq8.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:47 +0000 UTC Normal Pod simple-prod-query-6569c74cf6-rhtq8.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:47 +0000 UTC Normal Pod simple-prod-query-6569c74cf6-rhtq8.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:47 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6569c74cf6 SuccessfulDelete Deleted pod: simple-prod-query-6569c74cf6-rhtq8 replicaset-controller logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:47 +0000 UTC Normal Pod simple-prod-query-7d5467f79b-dvql9 Binding Scheduled Successfully assigned kuttl-test-relaxing-gator/simple-prod-query-7d5467f79b-dvql9 to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:47 +0000 UTC Normal ReplicaSet.apps simple-prod-query-7d5467f79b SuccessfulCreate Created pod: simple-prod-query-7d5467f79b-dvql9 replicaset-controller logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:47 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-6569c74cf6 to 0 from 1 deployment-controller logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:47 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-7d5467f79b to 1 deployment-controller logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:48 +0000 UTC Normal Pod simple-prod-query-7d5467f79b-dvql9 AddedInterface Add eth0 [10.128.2.66/23] from ovn-kubernetes multus logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:48 +0000 UTC Normal Pod simple-prod-query-7d5467f79b-dvql9.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:48 +0000 UTC Normal Pod simple-prod-query-7d5467f79b-dvql9.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:48 +0000 UTC Normal Pod simple-prod-query-7d5467f79b-dvql9.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:48 +0000 UTC Normal Pod simple-prod-query-7d5467f79b-dvql9.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:48 +0000 UTC Normal Pod simple-prod-query-7d5467f79b-dvql9.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:48 +0000 UTC Normal Pod simple-prod-query-7d5467f79b-dvql9.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:48 +0000 UTC Normal Pod simple-prod-query-7d5467f79b-dvql9.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:48 +0000 UTC Normal Pod simple-prod-query-7d5467f79b-dvql9.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:48 +0000 UTC Normal Pod simple-prod-query-7d5467f79b-dvql9.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:52 +0000 UTC Normal Pod check-span-92lqn Binding Scheduled Successfully assigned kuttl-test-relaxing-gator/check-span-92lqn to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:52 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-92lqn job-controller logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:52 +0000 UTC Normal Pod report-span-56q45 Binding Scheduled Successfully assigned kuttl-test-relaxing-gator/report-span-56q45 to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:52 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-56q45 job-controller logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:53 +0000 UTC Normal Pod check-span-92lqn AddedInterface Add eth0 [10.129.2.50/23] from ovn-kubernetes multus logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:53 +0000 UTC Normal Pod check-span-92lqn.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:53 +0000 UTC Normal Pod report-span-56q45 AddedInterface Add eth0 [10.129.2.49/23] from ovn-kubernetes multus logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:53 +0000 UTC Normal Pod report-span-56q45.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:53 +0000 UTC Normal Pod report-span-56q45.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 844ms (844ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:53 +0000 UTC Normal Pod report-span-56q45.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:53 +0000 UTC Normal Pod report-span-56q45.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:55 +0000 UTC Normal Pod check-span-92lqn.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 2.314s (2.314s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:55 +0000 UTC Normal Pod check-span-92lqn.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:55 +0000 UTC Normal Pod check-span-92lqn.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:55 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:55 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:33:55 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:34:06 | examples-simple-prod | 2025-01-20 07:34:05 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:34:06 | examples-simple-prod | Deleting namespace: kuttl-test-relaxing-gator === CONT kuttl/harness/examples-with-sampling logger.go:42: 07:34:19 | examples-with-sampling | Creating namespace: kuttl-test-exact-urchin logger.go:42: 07:34:19 | examples-with-sampling/0-install | starting test step 0-install logger.go:42: 07:34:19 | examples-with-sampling/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 07:34:19 | examples-with-sampling/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 07:34:19 | examples-with-sampling/0-install | >>>> Creating namespace kuttl-test-exact-urchin logger.go:42: 07:34:19 | examples-with-sampling/0-install | kubectl create namespace kuttl-test-exact-urchin 2>&1 | grep -v "already exists" || true logger.go:42: 07:34:19 | examples-with-sampling/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-exact-urchin 2>&1 | grep -v "already exists" || true logger.go:42: 07:34:19 | examples-with-sampling/0-install | service/cassandra created logger.go:42: 07:34:19 | examples-with-sampling/0-install | statefulset.apps/cassandra created logger.go:42: 07:34:19 | examples-with-sampling/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 07:34:30 | examples-with-sampling/0-install | test step completed 0-install logger.go:42: 07:34:30 | examples-with-sampling/1-install | starting test step 1-install logger.go:42: 07:34:30 | examples-with-sampling/1-install | Jaeger:kuttl-test-exact-urchin/with-sampling created logger.go:42: 07:34:36 | examples-with-sampling/1-install | test step completed 1-install logger.go:42: 07:34:36 | examples-with-sampling/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:34:36 | examples-with-sampling/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-sampling /dev/null] logger.go:42: 07:34:37 | examples-with-sampling/2-smoke-test | Warning: resource jaegers/with-sampling is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:34:44 | examples-with-sampling/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:34:45 | examples-with-sampling/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:34:45 | examples-with-sampling/2-smoke-test | job.batch/report-span created logger.go:42: 07:34:45 | examples-with-sampling/2-smoke-test | job.batch/check-span created logger.go:42: 07:34:58 | examples-with-sampling/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:34:58 | examples-with-sampling/3- | starting test step 3- logger.go:42: 07:34:58 | examples-with-sampling/3- | test step completed 3- logger.go:42: 07:34:58 | examples-with-sampling | examples-with-sampling events from ns kuttl-test-exact-urchin: logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:19 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-exact-urchin/cassandra-0 to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:19 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:20 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.131.0.29/23] from ovn-kubernetes multus logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:20 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:24 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 3.692s (3.692s including waiting). Image size: 309817322 bytes. kubelet logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:24 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:24 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:24 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-exact-urchin/cassandra-1 to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:24 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:25 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.129.2.51/23] from ovn-kubernetes multus logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:25 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:28 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 3.657s (3.657s including waiting). Image size: 309817322 bytes. kubelet logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:28 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:28 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:34 +0000 UTC Normal Pod with-sampling-5b86c6d4cf-nzl2f Binding Scheduled Successfully assigned kuttl-test-exact-urchin/with-sampling-5b86c6d4cf-nzl2f to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:34 +0000 UTC Normal Pod with-sampling-5b86c6d4cf-nzl2f AddedInterface Add eth0 [10.131.0.30/23] from ovn-kubernetes multus logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:34 +0000 UTC Normal Pod with-sampling-5b86c6d4cf-nzl2f.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:34 +0000 UTC Normal Pod with-sampling-5b86c6d4cf-nzl2f.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:34 +0000 UTC Normal Pod with-sampling-5b86c6d4cf-nzl2f.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:34 +0000 UTC Normal Pod with-sampling-5b86c6d4cf-nzl2f.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:34 +0000 UTC Normal Pod with-sampling-5b86c6d4cf-nzl2f.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:34 +0000 UTC Normal Pod with-sampling-5b86c6d4cf-nzl2f.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:34 +0000 UTC Normal ReplicaSet.apps with-sampling-5b86c6d4cf SuccessfulCreate Created pod: with-sampling-5b86c6d4cf-nzl2f replicaset-controller logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:34 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-5b86c6d4cf to 1 deployment-controller logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:41 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled down replica set with-sampling-5b86c6d4cf to 0 from 1 deployment-controller logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:42 +0000 UTC Normal Pod with-sampling-5b86c6d4cf-nzl2f.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:42 +0000 UTC Normal Pod with-sampling-5b86c6d4cf-nzl2f.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:42 +0000 UTC Normal ReplicaSet.apps with-sampling-5b86c6d4cf SuccessfulDelete Deleted pod: with-sampling-5b86c6d4cf-nzl2f replicaset-controller logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:42 +0000 UTC Normal Pod with-sampling-fbbf9bc55-zrrv4 Binding Scheduled Successfully assigned kuttl-test-exact-urchin/with-sampling-fbbf9bc55-zrrv4 to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:42 +0000 UTC Normal ReplicaSet.apps with-sampling-fbbf9bc55 SuccessfulCreate Created pod: with-sampling-fbbf9bc55-zrrv4 replicaset-controller logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:42 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-fbbf9bc55 to 1 deployment-controller logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:43 +0000 UTC Normal Pod with-sampling-fbbf9bc55-zrrv4 AddedInterface Add eth0 [10.131.0.31/23] from ovn-kubernetes multus logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:43 +0000 UTC Normal Pod with-sampling-fbbf9bc55-zrrv4.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:43 +0000 UTC Normal Pod with-sampling-fbbf9bc55-zrrv4.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:43 +0000 UTC Normal Pod with-sampling-fbbf9bc55-zrrv4.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:43 +0000 UTC Normal Pod with-sampling-fbbf9bc55-zrrv4.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:43 +0000 UTC Normal Pod with-sampling-fbbf9bc55-zrrv4.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:43 +0000 UTC Normal Pod with-sampling-fbbf9bc55-zrrv4.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:45 +0000 UTC Normal Pod check-span-kzxxd Binding Scheduled Successfully assigned kuttl-test-exact-urchin/check-span-kzxxd to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:45 +0000 UTC Normal Pod check-span-kzxxd AddedInterface Add eth0 [10.129.2.52/23] from ovn-kubernetes multus logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:45 +0000 UTC Normal Pod check-span-kzxxd.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:45 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-kzxxd job-controller logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:45 +0000 UTC Normal Pod report-span-lh8gm Binding Scheduled Successfully assigned kuttl-test-exact-urchin/report-span-lh8gm to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:45 +0000 UTC Normal Pod report-span-lh8gm AddedInterface Add eth0 [10.128.2.67/23] from ovn-kubernetes multus logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:45 +0000 UTC Normal Pod report-span-lh8gm.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:45 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-lh8gm job-controller logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:47 +0000 UTC Normal Pod check-span-kzxxd.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.301s (1.301s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:47 +0000 UTC Normal Pod check-span-kzxxd.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:47 +0000 UTC Normal Pod check-span-kzxxd.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:47 +0000 UTC Normal Pod report-span-lh8gm.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.184s (1.184s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:47 +0000 UTC Normal Pod report-span-lh8gm.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:47 +0000 UTC Normal Pod report-span-lh8gm.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:34:58 | examples-with-sampling | 2025-01-20 07:34:57 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:34:58 | examples-with-sampling | Deleting namespace: kuttl-test-exact-urchin === CONT kuttl/harness/examples-with-cassandra logger.go:42: 07:35:17 | examples-with-cassandra | Creating namespace: kuttl-test-correct-monster logger.go:42: 07:35:17 | examples-with-cassandra/0-install | starting test step 0-install logger.go:42: 07:35:17 | examples-with-cassandra/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 07:35:17 | examples-with-cassandra/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 07:35:17 | examples-with-cassandra/0-install | >>>> Creating namespace kuttl-test-correct-monster logger.go:42: 07:35:17 | examples-with-cassandra/0-install | kubectl create namespace kuttl-test-correct-monster 2>&1 | grep -v "already exists" || true logger.go:42: 07:35:17 | examples-with-cassandra/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-correct-monster 2>&1 | grep -v "already exists" || true logger.go:42: 07:35:17 | examples-with-cassandra/0-install | service/cassandra created logger.go:42: 07:35:17 | examples-with-cassandra/0-install | statefulset.apps/cassandra created logger.go:42: 07:35:17 | examples-with-cassandra/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 07:35:21 | examples-with-cassandra/0-install | test step completed 0-install logger.go:42: 07:35:21 | examples-with-cassandra/1-install | starting test step 1-install logger.go:42: 07:35:21 | examples-with-cassandra/1-install | Jaeger:kuttl-test-correct-monster/with-cassandra created logger.go:42: 07:35:39 | examples-with-cassandra/1-install | test step completed 1-install logger.go:42: 07:35:39 | examples-with-cassandra/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:35:39 | examples-with-cassandra/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-cassandra /dev/null] logger.go:42: 07:35:40 | examples-with-cassandra/2-smoke-test | Warning: resource jaegers/with-cassandra is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:35:47 | examples-with-cassandra/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:35:47 | examples-with-cassandra/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:35:48 | examples-with-cassandra/2-smoke-test | job.batch/report-span created logger.go:42: 07:35:48 | examples-with-cassandra/2-smoke-test | job.batch/check-span created logger.go:42: 07:36:01 | examples-with-cassandra/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:36:01 | examples-with-cassandra | examples-with-cassandra events from ns kuttl-test-correct-monster: logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:17 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-correct-monster/cassandra-0 to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:17 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:18 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.131.0.32/23] from ovn-kubernetes multus logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:18 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:18 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:18 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:19 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-correct-monster/cassandra-1 to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:19 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:20 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.129.2.53/23] from ovn-kubernetes multus logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:20 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:20 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:20 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:24 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-mxhmv Binding Scheduled Successfully assigned kuttl-test-correct-monster/with-cassandra-cassandra-schema-job-mxhmv to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:24 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job SuccessfulCreate Created pod: with-cassandra-cassandra-schema-job-mxhmv job-controller logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:25 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-mxhmv AddedInterface Add eth0 [10.128.2.68/23] from ovn-kubernetes multus logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:25 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-mxhmv.spec.containers{with-cassandra-cassandra-schema-job} Pulling Pulling image "jaegertracing/jaeger-cassandra-schema:1.57.0" kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:29 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-mxhmv.spec.containers{with-cassandra-cassandra-schema-job} Pulled Successfully pulled image "jaegertracing/jaeger-cassandra-schema:1.57.0" in 4.108s (4.108s including waiting). Image size: 381891613 bytes. kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:29 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-mxhmv.spec.containers{with-cassandra-cassandra-schema-job} Created Created container with-cassandra-cassandra-schema-job kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:29 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-mxhmv.spec.containers{with-cassandra-cassandra-schema-job} Started Started container with-cassandra-cassandra-schema-job kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:36 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job Completed Job completed job-controller logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:37 +0000 UTC Normal Pod with-cassandra-86b4fdf45c-pf5dh Binding Scheduled Successfully assigned kuttl-test-correct-monster/with-cassandra-86b4fdf45c-pf5dh to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:37 +0000 UTC Normal Pod with-cassandra-86b4fdf45c-pf5dh AddedInterface Add eth0 [10.131.0.33/23] from ovn-kubernetes multus logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:37 +0000 UTC Normal Pod with-cassandra-86b4fdf45c-pf5dh.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:37 +0000 UTC Normal Pod with-cassandra-86b4fdf45c-pf5dh.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:37 +0000 UTC Normal Pod with-cassandra-86b4fdf45c-pf5dh.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:37 +0000 UTC Normal Pod with-cassandra-86b4fdf45c-pf5dh.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:37 +0000 UTC Normal Pod with-cassandra-86b4fdf45c-pf5dh.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:37 +0000 UTC Normal Pod with-cassandra-86b4fdf45c-pf5dh.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:37 +0000 UTC Normal ReplicaSet.apps with-cassandra-86b4fdf45c SuccessfulCreate Created pod: with-cassandra-86b4fdf45c-pf5dh replicaset-controller logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:37 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-86b4fdf45c to 1 deployment-controller logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:43 +0000 UTC Normal Pod with-cassandra-86b4fdf45c-pf5dh.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:43 +0000 UTC Normal Pod with-cassandra-86b4fdf45c-pf5dh.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:43 +0000 UTC Normal ReplicaSet.apps with-cassandra-86b4fdf45c SuccessfulDelete Deleted pod: with-cassandra-86b4fdf45c-pf5dh replicaset-controller logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:43 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled down replica set with-cassandra-86b4fdf45c to 0 from 1 deployment-controller logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:44 +0000 UTC Normal Pod with-cassandra-7f5565cdf7-7p5q5 Binding Scheduled Successfully assigned kuttl-test-correct-monster/with-cassandra-7f5565cdf7-7p5q5 to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:44 +0000 UTC Normal ReplicaSet.apps with-cassandra-7f5565cdf7 SuccessfulCreate Created pod: with-cassandra-7f5565cdf7-7p5q5 replicaset-controller logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:44 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-7f5565cdf7 to 1 deployment-controller logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:45 +0000 UTC Normal Pod with-cassandra-7f5565cdf7-7p5q5 AddedInterface Add eth0 [10.131.0.34/23] from ovn-kubernetes multus logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:45 +0000 UTC Normal Pod with-cassandra-7f5565cdf7-7p5q5.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:45 +0000 UTC Normal Pod with-cassandra-7f5565cdf7-7p5q5.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:45 +0000 UTC Normal Pod with-cassandra-7f5565cdf7-7p5q5.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:45 +0000 UTC Normal Pod with-cassandra-7f5565cdf7-7p5q5.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:45 +0000 UTC Normal Pod with-cassandra-7f5565cdf7-7p5q5.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:45 +0000 UTC Normal Pod with-cassandra-7f5565cdf7-7p5q5.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:48 +0000 UTC Normal Pod check-span-ddcdf Binding Scheduled Successfully assigned kuttl-test-correct-monster/check-span-ddcdf to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:48 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-ddcdf job-controller logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:48 +0000 UTC Normal Pod report-span-8jmw8 Binding Scheduled Successfully assigned kuttl-test-correct-monster/report-span-8jmw8 to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:48 +0000 UTC Normal Pod report-span-8jmw8 AddedInterface Add eth0 [10.128.2.69/23] from ovn-kubernetes multus logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:48 +0000 UTC Normal Pod report-span-8jmw8.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:48 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-8jmw8 job-controller logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:49 +0000 UTC Normal Pod check-span-ddcdf AddedInterface Add eth0 [10.129.2.54/23] from ovn-kubernetes multus logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:49 +0000 UTC Normal Pod check-span-ddcdf.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:49 +0000 UTC Normal Pod report-span-8jmw8.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.183s (1.183s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:49 +0000 UTC Normal Pod report-span-8jmw8.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:49 +0000 UTC Normal Pod report-span-8jmw8.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:50 +0000 UTC Normal Pod check-span-ddcdf.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.49s (1.49s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:50 +0000 UTC Normal Pod check-span-ddcdf.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:35:50 +0000 UTC Normal Pod check-span-ddcdf.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:36:01 | examples-with-cassandra | 2025-01-20 07:36:01 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:36:01 | examples-with-cassandra | Deleting namespace: kuttl-test-correct-monster === CONT kuttl/harness/examples-with-badger logger.go:42: 07:36:13 | examples-with-badger | Creating namespace: kuttl-test-novel-ant logger.go:42: 07:36:14 | examples-with-badger/0-install | starting test step 0-install logger.go:42: 07:36:14 | examples-with-badger/0-install | Jaeger:kuttl-test-novel-ant/with-badger created logger.go:42: 07:36:20 | examples-with-badger/0-install | test step completed 0-install logger.go:42: 07:36:20 | examples-with-badger/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:36:20 | examples-with-badger/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-badger /dev/null] logger.go:42: 07:36:21 | examples-with-badger/1-smoke-test | Warning: resource jaegers/with-badger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:36:28 | examples-with-badger/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:36:28 | examples-with-badger/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:36:28 | examples-with-badger/1-smoke-test | job.batch/report-span created logger.go:42: 07:36:28 | examples-with-badger/1-smoke-test | job.batch/check-span created logger.go:42: 07:36:41 | examples-with-badger/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:36:41 | examples-with-badger | examples-with-badger events from ns kuttl-test-novel-ant: logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:18 +0000 UTC Normal Pod with-badger-5b8f489966-ltr65 Binding Scheduled Successfully assigned kuttl-test-novel-ant/with-badger-5b8f489966-ltr65 to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:18 +0000 UTC Normal Pod with-badger-5b8f489966-ltr65 AddedInterface Add eth0 [10.131.0.35/23] from ovn-kubernetes multus logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:18 +0000 UTC Normal Pod with-badger-5b8f489966-ltr65.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:18 +0000 UTC Normal Pod with-badger-5b8f489966-ltr65.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:18 +0000 UTC Normal Pod with-badger-5b8f489966-ltr65.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:18 +0000 UTC Normal Pod with-badger-5b8f489966-ltr65.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:18 +0000 UTC Normal Pod with-badger-5b8f489966-ltr65.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:18 +0000 UTC Normal Pod with-badger-5b8f489966-ltr65.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:18 +0000 UTC Normal ReplicaSet.apps with-badger-5b8f489966 SuccessfulCreate Created pod: with-badger-5b8f489966-ltr65 replicaset-controller logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:18 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-5b8f489966 to 1 deployment-controller logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:24 +0000 UTC Normal Pod with-badger-5b8f489966-ltr65.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:24 +0000 UTC Normal Pod with-badger-5b8f489966-ltr65.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:24 +0000 UTC Normal ReplicaSet.apps with-badger-5b8f489966 SuccessfulDelete Deleted pod: with-badger-5b8f489966-ltr65 replicaset-controller logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:24 +0000 UTC Normal Pod with-badger-5db86654f7-4s2wm Binding Scheduled Successfully assigned kuttl-test-novel-ant/with-badger-5db86654f7-4s2wm to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:24 +0000 UTC Normal ReplicaSet.apps with-badger-5db86654f7 SuccessfulCreate Created pod: with-badger-5db86654f7-4s2wm replicaset-controller logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:24 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled down replica set with-badger-5b8f489966 to 0 from 1 deployment-controller logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:24 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-5db86654f7 to 1 deployment-controller logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:25 +0000 UTC Normal Pod with-badger-5db86654f7-4s2wm AddedInterface Add eth0 [10.131.0.36/23] from ovn-kubernetes multus logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:25 +0000 UTC Normal Pod with-badger-5db86654f7-4s2wm.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:25 +0000 UTC Normal Pod with-badger-5db86654f7-4s2wm.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:25 +0000 UTC Normal Pod with-badger-5db86654f7-4s2wm.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:25 +0000 UTC Normal Pod with-badger-5db86654f7-4s2wm.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:25 +0000 UTC Normal Pod with-badger-5db86654f7-4s2wm.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:25 +0000 UTC Normal Pod with-badger-5db86654f7-4s2wm.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:28 +0000 UTC Normal Pod check-span-tglds Binding Scheduled Successfully assigned kuttl-test-novel-ant/check-span-tglds to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:28 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-tglds job-controller logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:28 +0000 UTC Normal Pod report-span-6ldqv Binding Scheduled Successfully assigned kuttl-test-novel-ant/report-span-6ldqv to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:28 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-6ldqv job-controller logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:29 +0000 UTC Normal Pod check-span-tglds AddedInterface Add eth0 [10.128.2.70/23] from ovn-kubernetes multus logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:29 +0000 UTC Normal Pod check-span-tglds.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:29 +0000 UTC Normal Pod report-span-6ldqv AddedInterface Add eth0 [10.129.2.55/23] from ovn-kubernetes multus logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:29 +0000 UTC Normal Pod report-span-6ldqv.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:29 +0000 UTC Normal Pod report-span-6ldqv.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 586ms (586ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:30 +0000 UTC Normal Pod check-span-tglds.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 896ms (896ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:30 +0000 UTC Normal Pod check-span-tglds.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:30 +0000 UTC Normal Pod check-span-tglds.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:30 +0000 UTC Normal Pod report-span-6ldqv.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:30 +0000 UTC Normal Pod report-span-6ldqv.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:36:41 | examples-with-badger | 2025-01-20 07:36:41 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:36:41 | examples-with-badger | Deleting namespace: kuttl-test-novel-ant === CONT kuttl/harness/examples-simplest logger.go:42: 07:36:47 | examples-simplest | Creating namespace: kuttl-test-precious-goose logger.go:42: 07:36:47 | examples-simplest/0-install | starting test step 0-install logger.go:42: 07:36:47 | examples-simplest/0-install | Jaeger:kuttl-test-precious-goose/simplest created logger.go:42: 07:36:55 | examples-simplest/0-install | test step completed 0-install logger.go:42: 07:36:55 | examples-simplest/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:36:55 | examples-simplest/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 07:36:56 | examples-simplest/1-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:37:02 | examples-simplest/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:37:03 | examples-simplest/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:37:03 | examples-simplest/1-smoke-test | job.batch/report-span created logger.go:42: 07:37:03 | examples-simplest/1-smoke-test | job.batch/check-span created logger.go:42: 07:37:17 | examples-simplest/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:37:17 | examples-simplest | examples-simplest events from ns kuttl-test-precious-goose: logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:36:51 +0000 UTC Normal Pod simplest-58f657fdb7-z7qkr Binding Scheduled Successfully assigned kuttl-test-precious-goose/simplest-58f657fdb7-z7qkr to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:36:51 +0000 UTC Normal ReplicaSet.apps simplest-58f657fdb7 SuccessfulCreate Created pod: simplest-58f657fdb7-z7qkr replicaset-controller logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:36:51 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-58f657fdb7 to 1 deployment-controller logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:36:52 +0000 UTC Warning Pod simplest-58f657fdb7-z7qkr FailedMount MountVolume.SetUp failed for volume "simplest-collector-tls-config-volume" : secret "simplest-collector-headless-tls" not found kubelet logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:36:53 +0000 UTC Normal Pod simplest-58f657fdb7-z7qkr AddedInterface Add eth0 [10.131.0.37/23] from ovn-kubernetes multus logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:36:53 +0000 UTC Normal Pod simplest-58f657fdb7-z7qkr.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:36:53 +0000 UTC Normal Pod simplest-58f657fdb7-z7qkr.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:36:53 +0000 UTC Normal Pod simplest-58f657fdb7-z7qkr.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:36:53 +0000 UTC Normal Pod simplest-58f657fdb7-z7qkr.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:36:53 +0000 UTC Normal Pod simplest-58f657fdb7-z7qkr.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:36:53 +0000 UTC Normal Pod simplest-58f657fdb7-z7qkr.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:36:59 +0000 UTC Normal Pod simplest-58f657fdb7-z7qkr.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:36:59 +0000 UTC Normal Pod simplest-58f657fdb7-z7qkr.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:36:59 +0000 UTC Normal ReplicaSet.apps simplest-58f657fdb7 SuccessfulDelete Deleted pod: simplest-58f657fdb7-z7qkr replicaset-controller logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:36:59 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-58f657fdb7 to 0 from 1 deployment-controller logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:37:00 +0000 UTC Normal Pod simplest-5d56cbf474-64jsj Binding Scheduled Successfully assigned kuttl-test-precious-goose/simplest-5d56cbf474-64jsj to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:37:00 +0000 UTC Normal ReplicaSet.apps simplest-5d56cbf474 SuccessfulCreate Created pod: simplest-5d56cbf474-64jsj replicaset-controller logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:37:00 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-5d56cbf474 to 1 deployment-controller logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:37:01 +0000 UTC Normal Pod simplest-5d56cbf474-64jsj AddedInterface Add eth0 [10.131.0.38/23] from ovn-kubernetes multus logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:37:01 +0000 UTC Normal Pod simplest-5d56cbf474-64jsj.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:37:01 +0000 UTC Normal Pod simplest-5d56cbf474-64jsj.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:37:01 +0000 UTC Normal Pod simplest-5d56cbf474-64jsj.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:37:01 +0000 UTC Normal Pod simplest-5d56cbf474-64jsj.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:37:01 +0000 UTC Normal Pod simplest-5d56cbf474-64jsj.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:37:01 +0000 UTC Normal Pod simplest-5d56cbf474-64jsj.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:37:03 +0000 UTC Normal Pod check-span-t92jk Binding Scheduled Successfully assigned kuttl-test-precious-goose/check-span-t92jk to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:37:03 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-t92jk job-controller logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:37:03 +0000 UTC Normal Pod report-span-xbmrh Binding Scheduled Successfully assigned kuttl-test-precious-goose/report-span-xbmrh to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:37:03 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-xbmrh job-controller logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:37:04 +0000 UTC Normal Pod check-span-t92jk AddedInterface Add eth0 [10.128.2.71/23] from ovn-kubernetes multus logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:37:04 +0000 UTC Normal Pod check-span-t92jk.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:37:04 +0000 UTC Normal Pod report-span-xbmrh AddedInterface Add eth0 [10.129.2.56/23] from ovn-kubernetes multus logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:37:04 +0000 UTC Normal Pod report-span-xbmrh.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:37:06 +0000 UTC Normal Pod check-span-t92jk.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.762s (1.762s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:37:06 +0000 UTC Normal Pod check-span-t92jk.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:37:06 +0000 UTC Normal Pod check-span-t92jk.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:37:09 +0000 UTC Normal Pod report-span-xbmrh.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 5.014s (5.014s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:37:09 +0000 UTC Normal Pod report-span-xbmrh.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:37:09 +0000 UTC Normal Pod report-span-xbmrh.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:37:17 | examples-simplest | 2025-01-20 07:37:17 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:37:17 | examples-simplest | Deleting namespace: kuttl-test-precious-goose === CONT kuttl/harness/examples-simple-prod-with-volumes logger.go:42: 07:37:24 | examples-simple-prod-with-volumes | Ignoring 03-check-volume.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:37:24 | examples-simple-prod-with-volumes | Creating namespace: kuttl-test-legible-reptile logger.go:42: 07:37:24 | examples-simple-prod-with-volumes/1-install | starting test step 1-install logger.go:42: 07:37:24 | examples-simple-prod-with-volumes/1-install | Jaeger:kuttl-test-legible-reptile/simple-prod created logger.go:42: 07:38:01 | examples-simple-prod-with-volumes/1-install | test step completed 1-install logger.go:42: 07:38:01 | examples-simple-prod-with-volumes/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:38:01 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:38:03 | examples-simple-prod-with-volumes/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:38:09 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:38:10 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:38:10 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/report-span created logger.go:42: 07:38:10 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/check-span created logger.go:42: 07:38:22 | examples-simple-prod-with-volumes/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:38:22 | examples-simple-prod-with-volumes/3-check-volume | starting test step 3-check-volume logger.go:42: 07:38:22 | examples-simple-prod-with-volumes/3-check-volume | running command: [sh -c kubectl exec $(kubectl get pods -n $NAMESPACE -l app=jaeger -l app.kubernetes.io/component=collector -o yaml | /tmp/jaeger-tests/bin/yq e '.items[0].metadata.name') -n $NAMESPACE -- ls /usr/share/elasticsearch/data] logger.go:42: 07:38:23 | examples-simple-prod-with-volumes/3-check-volume | test step completed 3-check-volume logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | examples-simple-prod-with-volumes events from ns kuttl-test-legible-reptile: logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:31 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestlegiblereptilesimpleprod-1-586fd4cbd7 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestlegiblereptilesimpleprod-1-586fvj78q replicaset-controller logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:31 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestlegiblereptilesimpleprod-1-586fvj78q Binding Scheduled Successfully assigned kuttl-test-legible-reptile/elasticsearch-cdm-kuttltestlegiblereptilesimpleprod-1-586fvj78q to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:31 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestlegiblereptilesimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestlegiblereptilesimpleprod-1-586fd4cbd7 to 1 deployment-controller logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestlegiblereptilesimpleprod-1-586fvj78q AddedInterface Add eth0 [10.131.0.39/23] from ovn-kubernetes multus logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestlegiblereptilesimpleprod-1-586fvj78q.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestlegiblereptilesimpleprod-1-586fvj78q.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestlegiblereptilesimpleprod-1-586fvj78q.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestlegiblereptilesimpleprod-1-586fvj78q.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestlegiblereptilesimpleprod-1-586fvj78q.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestlegiblereptilesimpleprod-1-586fvj78q.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:43 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestlegiblereptilesimpleprod-1-586fvj78q.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:48 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestlegiblereptilesimpleprod-1-586fvj78q.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:58 +0000 UTC Normal Pod simple-prod-collector-78988bd974-jgxfj Binding Scheduled Successfully assigned kuttl-test-legible-reptile/simple-prod-collector-78988bd974-jgxfj to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:58 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-78988bd974 SuccessfulCreate Created pod: simple-prod-collector-78988bd974-jgxfj replicaset-controller logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:58 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-78988bd974 to 1 deployment-controller logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:58 +0000 UTC Normal Pod simple-prod-query-6c6dcbcdff-mqtrj Binding Scheduled Successfully assigned kuttl-test-legible-reptile/simple-prod-query-6c6dcbcdff-mqtrj to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:58 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6c6dcbcdff SuccessfulCreate Created pod: simple-prod-query-6c6dcbcdff-mqtrj replicaset-controller logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:58 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-6c6dcbcdff to 1 deployment-controller logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:59 +0000 UTC Normal Pod simple-prod-collector-78988bd974-jgxfj AddedInterface Add eth0 [10.129.2.57/23] from ovn-kubernetes multus logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:59 +0000 UTC Normal Pod simple-prod-collector-78988bd974-jgxfj.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:59 +0000 UTC Normal Pod simple-prod-collector-78988bd974-jgxfj.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:59 +0000 UTC Normal Pod simple-prod-collector-78988bd974-jgxfj.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:59 +0000 UTC Normal Pod simple-prod-query-6c6dcbcdff-mqtrj AddedInterface Add eth0 [10.128.2.72/23] from ovn-kubernetes multus logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:59 +0000 UTC Normal Pod simple-prod-query-6c6dcbcdff-mqtrj.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:59 +0000 UTC Normal Pod simple-prod-query-6c6dcbcdff-mqtrj.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:59 +0000 UTC Normal Pod simple-prod-query-6c6dcbcdff-mqtrj.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:59 +0000 UTC Normal Pod simple-prod-query-6c6dcbcdff-mqtrj.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:59 +0000 UTC Normal Pod simple-prod-query-6c6dcbcdff-mqtrj.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:59 +0000 UTC Normal Pod simple-prod-query-6c6dcbcdff-mqtrj.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:59 +0000 UTC Normal Pod simple-prod-query-6c6dcbcdff-mqtrj.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:59 +0000 UTC Normal Pod simple-prod-query-6c6dcbcdff-mqtrj.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:37:59 +0000 UTC Normal Pod simple-prod-query-6c6dcbcdff-mqtrj.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:04 +0000 UTC Normal Pod simple-prod-query-6c6dcbcdff-mqtrj.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:04 +0000 UTC Normal Pod simple-prod-query-6c6dcbcdff-mqtrj.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:04 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-6c6dcbcdff to 0 from 1 deployment-controller logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:05 +0000 UTC Normal Pod simple-prod-query-6c6dcbcdff-mqtrj.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:05 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6c6dcbcdff SuccessfulDelete Deleted pod: simple-prod-query-6c6dcbcdff-mqtrj replicaset-controller logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:05 +0000 UTC Normal Pod simple-prod-query-86d9ccf76c-6fx2j Binding Scheduled Successfully assigned kuttl-test-legible-reptile/simple-prod-query-86d9ccf76c-6fx2j to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:05 +0000 UTC Normal ReplicaSet.apps simple-prod-query-86d9ccf76c SuccessfulCreate Created pod: simple-prod-query-86d9ccf76c-6fx2j replicaset-controller logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:05 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-86d9ccf76c to 1 deployment-controller logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:06 +0000 UTC Normal Pod simple-prod-query-86d9ccf76c-6fx2j AddedInterface Add eth0 [10.128.2.73/23] from ovn-kubernetes multus logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:06 +0000 UTC Normal Pod simple-prod-query-86d9ccf76c-6fx2j.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:06 +0000 UTC Normal Pod simple-prod-query-86d9ccf76c-6fx2j.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:06 +0000 UTC Normal Pod simple-prod-query-86d9ccf76c-6fx2j.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:06 +0000 UTC Normal Pod simple-prod-query-86d9ccf76c-6fx2j.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:06 +0000 UTC Normal Pod simple-prod-query-86d9ccf76c-6fx2j.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:06 +0000 UTC Normal Pod simple-prod-query-86d9ccf76c-6fx2j.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:06 +0000 UTC Normal Pod simple-prod-query-86d9ccf76c-6fx2j.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:06 +0000 UTC Normal Pod simple-prod-query-86d9ccf76c-6fx2j.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:06 +0000 UTC Normal Pod simple-prod-query-86d9ccf76c-6fx2j.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:10 +0000 UTC Normal Pod check-span-q54h5 Binding Scheduled Successfully assigned kuttl-test-legible-reptile/check-span-q54h5 to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:10 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-q54h5 job-controller logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:10 +0000 UTC Normal Pod report-span-c56px Binding Scheduled Successfully assigned kuttl-test-legible-reptile/report-span-c56px to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:10 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-c56px job-controller logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:11 +0000 UTC Normal Pod check-span-q54h5 AddedInterface Add eth0 [10.129.2.59/23] from ovn-kubernetes multus logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:11 +0000 UTC Normal Pod check-span-q54h5.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:11 +0000 UTC Normal Pod check-span-q54h5.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 761ms (761ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:11 +0000 UTC Normal Pod check-span-q54h5.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:11 +0000 UTC Normal Pod check-span-q54h5.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:11 +0000 UTC Normal Pod report-span-c56px AddedInterface Add eth0 [10.129.2.58/23] from ovn-kubernetes multus logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:11 +0000 UTC Normal Pod report-span-c56px.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:11 +0000 UTC Normal Pod report-span-c56px.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 596ms (596ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:11 +0000 UTC Normal Pod report-span-c56px.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:11 +0000 UTC Normal Pod report-span-c56px.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:13 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:13 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:13 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | 2025-01-20 07:38:22 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:38:23 | examples-simple-prod-with-volumes | Deleting namespace: kuttl-test-legible-reptile === CONT kuttl/harness/examples-business-application-injected-sidecar logger.go:42: 07:38:35 | examples-business-application-injected-sidecar | Creating namespace: kuttl-test-smart-skylark logger.go:42: 07:38:35 | examples-business-application-injected-sidecar/0-install | starting test step 0-install logger.go:42: 07:38:35 | examples-business-application-injected-sidecar/0-install | Deployment:kuttl-test-smart-skylark/myapp created logger.go:42: 07:38:35 | examples-business-application-injected-sidecar/0-install | test step completed 0-install logger.go:42: 07:38:35 | examples-business-application-injected-sidecar/1-install | starting test step 1-install logger.go:42: 07:38:35 | examples-business-application-injected-sidecar/1-install | Jaeger:kuttl-test-smart-skylark/simplest created logger.go:42: 07:38:46 | examples-business-application-injected-sidecar/1-install | test step completed 1-install logger.go:42: 07:38:46 | examples-business-application-injected-sidecar/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:38:46 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 07:38:48 | examples-business-application-injected-sidecar/2-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:38:54 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:38:55 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:38:55 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/report-span created logger.go:42: 07:38:55 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/check-span created logger.go:42: 07:39:07 | examples-business-application-injected-sidecar/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | examples-business-application-injected-sidecar events from ns kuttl-test-smart-skylark: logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:35 +0000 UTC Normal Pod myapp-fc54f78d8-6xgss Binding Scheduled Successfully assigned kuttl-test-smart-skylark/myapp-fc54f78d8-6xgss to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:35 +0000 UTC Normal ReplicaSet.apps myapp-fc54f78d8 SuccessfulCreate Created pod: myapp-fc54f78d8-6xgss replicaset-controller logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:35 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-fc54f78d8 to 1 deployment-controller logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:36 +0000 UTC Normal Pod myapp-59c4767f46-lbms8 Binding Scheduled Successfully assigned kuttl-test-smart-skylark/myapp-59c4767f46-lbms8 to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:36 +0000 UTC Normal ReplicaSet.apps myapp-59c4767f46 SuccessfulCreate Created pod: myapp-59c4767f46-lbms8 replicaset-controller logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:36 +0000 UTC Normal Pod myapp-fc54f78d8-6xgss AddedInterface Add eth0 [10.131.0.40/23] from ovn-kubernetes multus logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:36 +0000 UTC Normal Pod myapp-fc54f78d8-6xgss.spec.containers{myapp} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:36 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-59c4767f46 to 1 deployment-controller logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:37 +0000 UTC Warning Pod myapp-59c4767f46-lbms8 FailedMount MountVolume.SetUp failed for volume "simplest-trusted-ca" : configmap "simplest-trusted-ca" not found kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:37 +0000 UTC Warning Pod myapp-59c4767f46-lbms8 FailedMount MountVolume.SetUp failed for volume "simplest-service-ca" : configmap "simplest-service-ca" not found kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:39 +0000 UTC Normal Pod myapp-fc54f78d8-6xgss.spec.containers{myapp} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 3.334s (3.334s including waiting). Image size: 282912835 bytes. kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:39 +0000 UTC Normal Pod myapp-fc54f78d8-6xgss.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:39 +0000 UTC Normal Pod myapp-fc54f78d8-6xgss.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:44 +0000 UTC Normal Pod simplest-6849cc764c-v66q6 Binding Scheduled Successfully assigned kuttl-test-smart-skylark/simplest-6849cc764c-v66q6 to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:44 +0000 UTC Normal ReplicaSet.apps simplest-6849cc764c SuccessfulCreate Created pod: simplest-6849cc764c-v66q6 replicaset-controller logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:44 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-6849cc764c to 1 deployment-controller logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:45 +0000 UTC Normal Pod simplest-6849cc764c-v66q6 AddedInterface Add eth0 [10.131.0.41/23] from ovn-kubernetes multus logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:45 +0000 UTC Normal Pod simplest-6849cc764c-v66q6.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:45 +0000 UTC Normal Pod simplest-6849cc764c-v66q6.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:45 +0000 UTC Normal Pod simplest-6849cc764c-v66q6.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:45 +0000 UTC Normal Pod simplest-6849cc764c-v66q6.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:45 +0000 UTC Normal Pod simplest-6849cc764c-v66q6.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:45 +0000 UTC Normal Pod simplest-6849cc764c-v66q6.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:47 +0000 UTC Warning Pod myapp-fc54f78d8-6xgss.spec.containers{myapp} Unhealthy Liveness probe failed: Get "http://10.131.0.40:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:50 +0000 UTC Normal Pod simplest-6849cc764c-v66q6.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:50 +0000 UTC Normal Pod simplest-6849cc764c-v66q6.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:50 +0000 UTC Normal ReplicaSet.apps simplest-6849cc764c SuccessfulDelete Deleted pod: simplest-6849cc764c-v66q6 replicaset-controller logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:50 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-6849cc764c to 0 from 1 deployment-controller logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:51 +0000 UTC Normal Pod simplest-76f6ccfb59-4mbzb Binding Scheduled Successfully assigned kuttl-test-smart-skylark/simplest-76f6ccfb59-4mbzb to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:51 +0000 UTC Normal Pod simplest-76f6ccfb59-4mbzb AddedInterface Add eth0 [10.131.0.42/23] from ovn-kubernetes multus logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:51 +0000 UTC Normal Pod simplest-76f6ccfb59-4mbzb.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:51 +0000 UTC Normal Pod simplest-76f6ccfb59-4mbzb.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:51 +0000 UTC Normal Pod simplest-76f6ccfb59-4mbzb.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:51 +0000 UTC Normal Pod simplest-76f6ccfb59-4mbzb.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:51 +0000 UTC Normal Pod simplest-76f6ccfb59-4mbzb.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:51 +0000 UTC Normal Pod simplest-76f6ccfb59-4mbzb.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:51 +0000 UTC Normal ReplicaSet.apps simplest-76f6ccfb59 SuccessfulCreate Created pod: simplest-76f6ccfb59-4mbzb replicaset-controller logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:51 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-76f6ccfb59 to 1 deployment-controller logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:53 +0000 UTC Normal Pod myapp-59c4767f46-lbms8 AddedInterface Add eth0 [10.129.2.60/23] from ovn-kubernetes multus logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:53 +0000 UTC Normal Pod myapp-59c4767f46-lbms8.spec.containers{myapp} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:55 +0000 UTC Normal Pod check-span-wg8gm Binding Scheduled Successfully assigned kuttl-test-smart-skylark/check-span-wg8gm to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:55 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-wg8gm job-controller logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:55 +0000 UTC Normal Pod report-span-rqxs2 Binding Scheduled Successfully assigned kuttl-test-smart-skylark/report-span-rqxs2 to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:55 +0000 UTC Normal Pod report-span-rqxs2 AddedInterface Add eth0 [10.128.2.74/23] from ovn-kubernetes multus logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:55 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-rqxs2 job-controller logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:56 +0000 UTC Normal Pod check-span-wg8gm AddedInterface Add eth0 [10.128.2.75/23] from ovn-kubernetes multus logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:56 +0000 UTC Normal Pod check-span-wg8gm.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:56 +0000 UTC Normal Pod check-span-wg8gm.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 561ms (561ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:56 +0000 UTC Normal Pod check-span-wg8gm.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:56 +0000 UTC Normal Pod check-span-wg8gm.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:56 +0000 UTC Normal Pod myapp-59c4767f46-lbms8.spec.containers{myapp} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 3.195s (3.195s including waiting). Image size: 282912835 bytes. kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:56 +0000 UTC Normal Pod myapp-59c4767f46-lbms8.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:56 +0000 UTC Normal Pod myapp-59c4767f46-lbms8.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:56 +0000 UTC Normal Pod myapp-59c4767f46-lbms8.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:56 +0000 UTC Normal Pod myapp-59c4767f46-lbms8.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:56 +0000 UTC Normal Pod myapp-59c4767f46-lbms8.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:56 +0000 UTC Normal Pod report-span-rqxs2.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:56 +0000 UTC Normal Pod report-span-rqxs2.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 598ms (598ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:56 +0000 UTC Normal Pod report-span-rqxs2.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:56 +0000 UTC Normal Pod report-span-rqxs2.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:57 +0000 UTC Normal Pod myapp-fc54f78d8-6xgss.spec.containers{myapp} Killing Stopping container myapp kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:57 +0000 UTC Normal ReplicaSet.apps myapp-fc54f78d8 SuccessfulDelete Deleted pod: myapp-fc54f78d8-6xgss replicaset-controller logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:38:57 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled down replica set myapp-fc54f78d8 to 0 from 1 deployment-controller logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:39:03 +0000 UTC Warning Pod myapp-59c4767f46-lbms8.spec.containers{myapp} Unhealthy Liveness probe failed: Get "http://10.129.2.60:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | 2025-01-20 07:39:07 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:39:07 | examples-business-application-injected-sidecar | Deleting namespace: kuttl-test-smart-skylark === CONT kuttl/harness/examples-service-types logger.go:42: 07:39:14 | examples-service-types | Creating namespace: kuttl-test-improved-bat logger.go:42: 07:39:14 | examples-service-types/0-install | starting test step 0-install logger.go:42: 07:39:14 | examples-service-types/0-install | Jaeger:kuttl-test-improved-bat/service-types created logger.go:42: 07:39:20 | examples-service-types/0-install | test step completed 0-install logger.go:42: 07:39:20 | examples-service-types/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:39:20 | examples-service-types/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE service-types /dev/null] logger.go:42: 07:39:22 | examples-service-types/1-smoke-test | Warning: resource jaegers/service-types is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:39:28 | examples-service-types/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://service-types-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:39:29 | examples-service-types/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:39:29 | examples-service-types/1-smoke-test | job.batch/report-span created logger.go:42: 07:39:29 | examples-service-types/1-smoke-test | job.batch/check-span created logger.go:42: 07:39:43 | examples-service-types/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:39:43 | examples-service-types/2- | starting test step 2- logger.go:42: 07:39:43 | examples-service-types/2- | test step completed 2- logger.go:42: 07:39:43 | examples-service-types | examples-service-types events from ns kuttl-test-improved-bat: logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:18 +0000 UTC Normal Pod service-types-565f8cd6-ldlnk Binding Scheduled Successfully assigned kuttl-test-improved-bat/service-types-565f8cd6-ldlnk to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:18 +0000 UTC Normal Pod service-types-565f8cd6-ldlnk AddedInterface Add eth0 [10.131.0.43/23] from ovn-kubernetes multus logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:18 +0000 UTC Normal Pod service-types-565f8cd6-ldlnk.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:18 +0000 UTC Normal ReplicaSet.apps service-types-565f8cd6 SuccessfulCreate Created pod: service-types-565f8cd6-ldlnk replicaset-controller logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:18 +0000 UTC Normal Service service-types-collector EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:18 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-565f8cd6 to 1 deployment-controller logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:19 +0000 UTC Normal Pod service-types-565f8cd6-ldlnk.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:19 +0000 UTC Normal Pod service-types-565f8cd6-ldlnk.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:19 +0000 UTC Normal Pod service-types-565f8cd6-ldlnk.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:19 +0000 UTC Normal Pod service-types-565f8cd6-ldlnk.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:19 +0000 UTC Normal Pod service-types-565f8cd6-ldlnk.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:21 +0000 UTC Normal Service service-types-collector EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:21 +0000 UTC Normal Service service-types-query EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:24 +0000 UTC Normal Service service-types-query EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:25 +0000 UTC Normal Pod service-types-565f8cd6-ldlnk.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:25 +0000 UTC Normal Pod service-types-565f8cd6-ldlnk.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:25 +0000 UTC Normal ReplicaSet.apps service-types-565f8cd6 SuccessfulDelete Deleted pod: service-types-565f8cd6-ldlnk replicaset-controller logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:25 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled down replica set service-types-565f8cd6 to 0 from 1 deployment-controller logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:26 +0000 UTC Normal Pod service-types-f6bd898b5-gjt5n Binding Scheduled Successfully assigned kuttl-test-improved-bat/service-types-f6bd898b5-gjt5n to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:26 +0000 UTC Normal Pod service-types-f6bd898b5-gjt5n AddedInterface Add eth0 [10.131.0.44/23] from ovn-kubernetes multus logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:26 +0000 UTC Normal Pod service-types-f6bd898b5-gjt5n.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:26 +0000 UTC Normal Pod service-types-f6bd898b5-gjt5n.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:26 +0000 UTC Normal Pod service-types-f6bd898b5-gjt5n.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:26 +0000 UTC Normal Pod service-types-f6bd898b5-gjt5n.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:26 +0000 UTC Normal ReplicaSet.apps service-types-f6bd898b5 SuccessfulCreate Created pod: service-types-f6bd898b5-gjt5n replicaset-controller logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:26 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-f6bd898b5 to 1 deployment-controller logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:27 +0000 UTC Normal Pod service-types-f6bd898b5-gjt5n.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:27 +0000 UTC Normal Pod service-types-f6bd898b5-gjt5n.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:29 +0000 UTC Normal Pod check-span-wrnfq Binding Scheduled Successfully assigned kuttl-test-improved-bat/check-span-wrnfq to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:29 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-wrnfq job-controller logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:29 +0000 UTC Normal Pod report-span-46h22 Binding Scheduled Successfully assigned kuttl-test-improved-bat/report-span-46h22 to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:29 +0000 UTC Normal Pod report-span-46h22 AddedInterface Add eth0 [10.129.2.61/23] from ovn-kubernetes multus logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:29 +0000 UTC Normal Pod report-span-46h22.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:29 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-46h22 job-controller logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:30 +0000 UTC Normal Pod check-span-wrnfq AddedInterface Add eth0 [10.128.2.76/23] from ovn-kubernetes multus logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:30 +0000 UTC Normal Pod check-span-wrnfq.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:32 +0000 UTC Normal Pod check-span-wrnfq.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 2.728s (2.728s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:32 +0000 UTC Normal Pod check-span-wrnfq.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:32 +0000 UTC Normal Pod check-span-wrnfq.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:33 +0000 UTC Normal Pod report-span-46h22.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 3.363s (3.363s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:33 +0000 UTC Normal Pod report-span-46h22.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:33 +0000 UTC Normal Pod report-span-46h22.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:39:43 | examples-service-types | 2025-01-20 07:39:43 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:39:43 | examples-service-types | Deleting namespace: kuttl-test-improved-bat === CONT kuttl/harness/examples-openshift-with-htpasswd logger.go:42: 07:40:11 | examples-openshift-with-htpasswd | Ignoring 00-install.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:40:11 | examples-openshift-with-htpasswd | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:40:11 | examples-openshift-with-htpasswd | Creating namespace: kuttl-test-safe-terrapin logger.go:42: 07:40:11 | examples-openshift-with-htpasswd/0-install | starting test step 0-install logger.go:42: 07:40:11 | examples-openshift-with-htpasswd/0-install | Secret:kuttl-test-safe-terrapin/htpasswd created logger.go:42: 07:40:11 | examples-openshift-with-htpasswd/0-install | test step completed 0-install logger.go:42: 07:40:11 | examples-openshift-with-htpasswd/1-install | starting test step 1-install logger.go:42: 07:40:11 | examples-openshift-with-htpasswd/1-install | Jaeger:kuttl-test-safe-terrapin/with-htpasswd created logger.go:42: 07:40:17 | examples-openshift-with-htpasswd/1-install | test step completed 1-install logger.go:42: 07:40:17 | examples-openshift-with-htpasswd/2-check-unsecured | starting test step 2-check-unsecured logger.go:42: 07:40:17 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [./ensure-ingress-host.sh] logger.go:42: 07:40:17 | examples-openshift-with-htpasswd/2-check-unsecured | Checking the Ingress host value was populated logger.go:42: 07:40:17 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 0 logger.go:42: 07:40:17 | examples-openshift-with-htpasswd/2-check-unsecured | Hostname is with-htpasswd-kuttl-test-safe-terrapin.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com logger.go:42: 07:40:17 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 07:40:18 | examples-openshift-with-htpasswd/2-check-unsecured | Checking an expected HTTP response logger.go:42: 07:40:18 | examples-openshift-with-htpasswd/2-check-unsecured | Running in OpenShift logger.go:42: 07:40:18 | examples-openshift-with-htpasswd/2-check-unsecured | Not using any secret logger.go:42: 07:40:18 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 1/30 the https://with-htpasswd-kuttl-test-safe-terrapin.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:40:18 | examples-openshift-with-htpasswd/2-check-unsecured | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:40:18 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 2/30 the https://with-htpasswd-kuttl-test-safe-terrapin.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:40:18 | examples-openshift-with-htpasswd/2-check-unsecured | HTTP response is 503. 403 expected. Waiting 10 s logger.go:42: 07:40:28 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 3/30 the https://with-htpasswd-kuttl-test-safe-terrapin.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:40:28 | examples-openshift-with-htpasswd/2-check-unsecured | curl response asserted properly logger.go:42: 07:40:28 | examples-openshift-with-htpasswd/2-check-unsecured | test step completed 2-check-unsecured logger.go:42: 07:40:28 | examples-openshift-with-htpasswd/3-check-unauthorized | starting test step 3-check-unauthorized logger.go:42: 07:40:28 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [./ensure-ingress-host.sh] logger.go:42: 07:40:28 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking the Ingress host value was populated logger.go:42: 07:40:28 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 0 logger.go:42: 07:40:28 | examples-openshift-with-htpasswd/3-check-unauthorized | Hostname is with-htpasswd-kuttl-test-safe-terrapin.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com logger.go:42: 07:40:28 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [sh -c JAEGER_USERNAME=wronguser JAEGER_PASSWORD=wrongpassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 07:40:28 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking an expected HTTP response logger.go:42: 07:40:28 | examples-openshift-with-htpasswd/3-check-unauthorized | Running in OpenShift logger.go:42: 07:40:28 | examples-openshift-with-htpasswd/3-check-unauthorized | Using Jaeger basic authentication logger.go:42: 07:40:28 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 1/30 the https://with-htpasswd-kuttl-test-safe-terrapin.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:40:28 | examples-openshift-with-htpasswd/3-check-unauthorized | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:40:28 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 2/30 the https://with-htpasswd-kuttl-test-safe-terrapin.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:40:28 | examples-openshift-with-htpasswd/3-check-unauthorized | curl response asserted properly logger.go:42: 07:40:28 | examples-openshift-with-htpasswd/3-check-unauthorized | test step completed 3-check-unauthorized logger.go:42: 07:40:28 | examples-openshift-with-htpasswd/4-check-authorized | starting test step 4-check-authorized logger.go:42: 07:40:28 | examples-openshift-with-htpasswd/4-check-authorized | running command: [./ensure-ingress-host.sh] logger.go:42: 07:40:28 | examples-openshift-with-htpasswd/4-check-authorized | Checking the Ingress host value was populated logger.go:42: 07:40:28 | examples-openshift-with-htpasswd/4-check-authorized | Try number 0 logger.go:42: 07:40:28 | examples-openshift-with-htpasswd/4-check-authorized | Hostname is with-htpasswd-kuttl-test-safe-terrapin.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com logger.go:42: 07:40:28 | examples-openshift-with-htpasswd/4-check-authorized | running command: [sh -c JAEGER_USERNAME=awesomeuser JAEGER_PASSWORD=awesomepassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE with-htpasswd] logger.go:42: 07:40:28 | examples-openshift-with-htpasswd/4-check-authorized | Checking an expected HTTP response logger.go:42: 07:40:28 | examples-openshift-with-htpasswd/4-check-authorized | Running in OpenShift logger.go:42: 07:40:28 | examples-openshift-with-htpasswd/4-check-authorized | Using Jaeger basic authentication logger.go:42: 07:40:28 | examples-openshift-with-htpasswd/4-check-authorized | Try number 1/30 the https://with-htpasswd-kuttl-test-safe-terrapin.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:40:28 | examples-openshift-with-htpasswd/4-check-authorized | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:40:28 | examples-openshift-with-htpasswd/4-check-authorized | Try number 2/30 the https://with-htpasswd-kuttl-test-safe-terrapin.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:40:29 | examples-openshift-with-htpasswd/4-check-authorized | curl response asserted properly logger.go:42: 07:40:29 | examples-openshift-with-htpasswd/4-check-authorized | test step completed 4-check-authorized logger.go:42: 07:40:29 | examples-openshift-with-htpasswd | examples-openshift-with-htpasswd events from ns kuttl-test-safe-terrapin: logger.go:42: 07:40:29 | examples-openshift-with-htpasswd | 2025-01-20 07:40:15 +0000 UTC Normal Pod with-htpasswd-775b66685-cs7wl Binding Scheduled Successfully assigned kuttl-test-safe-terrapin/with-htpasswd-775b66685-cs7wl to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:40:29 | examples-openshift-with-htpasswd | 2025-01-20 07:40:15 +0000 UTC Normal ReplicaSet.apps with-htpasswd-775b66685 SuccessfulCreate Created pod: with-htpasswd-775b66685-cs7wl replicaset-controller logger.go:42: 07:40:29 | examples-openshift-with-htpasswd | 2025-01-20 07:40:15 +0000 UTC Normal Deployment.apps with-htpasswd ScalingReplicaSet Scaled up replica set with-htpasswd-775b66685 to 1 deployment-controller logger.go:42: 07:40:29 | examples-openshift-with-htpasswd | 2025-01-20 07:40:16 +0000 UTC Normal Pod with-htpasswd-775b66685-cs7wl AddedInterface Add eth0 [10.131.0.45/23] from ovn-kubernetes multus logger.go:42: 07:40:29 | examples-openshift-with-htpasswd | 2025-01-20 07:40:16 +0000 UTC Normal Pod with-htpasswd-775b66685-cs7wl.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:40:29 | examples-openshift-with-htpasswd | 2025-01-20 07:40:16 +0000 UTC Normal Pod with-htpasswd-775b66685-cs7wl.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:40:29 | examples-openshift-with-htpasswd | 2025-01-20 07:40:16 +0000 UTC Normal Pod with-htpasswd-775b66685-cs7wl.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:40:29 | examples-openshift-with-htpasswd | 2025-01-20 07:40:16 +0000 UTC Normal Pod with-htpasswd-775b66685-cs7wl.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:40:29 | examples-openshift-with-htpasswd | 2025-01-20 07:40:16 +0000 UTC Normal Pod with-htpasswd-775b66685-cs7wl.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:40:29 | examples-openshift-with-htpasswd | 2025-01-20 07:40:16 +0000 UTC Normal Pod with-htpasswd-775b66685-cs7wl.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:40:29 | examples-openshift-with-htpasswd | Deleting namespace: kuttl-test-safe-terrapin === CONT kuttl/harness/examples-all-in-one-with-options logger.go:42: 07:40:35 | examples-all-in-one-with-options | Creating namespace: kuttl-test-enjoyed-mayfly logger.go:42: 07:40:35 | examples-all-in-one-with-options/0-install | starting test step 0-install logger.go:42: 07:40:35 | examples-all-in-one-with-options/0-install | Jaeger:kuttl-test-enjoyed-mayfly/my-jaeger created logger.go:42: 07:40:41 | examples-all-in-one-with-options/0-install | test step completed 0-install logger.go:42: 07:40:41 | examples-all-in-one-with-options/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:40:41 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:40:43 | examples-all-in-one-with-options/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:40:49 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443/jaeger MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:40:50 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:40:50 | examples-all-in-one-with-options/1-smoke-test | job.batch/report-span created logger.go:42: 07:40:50 | examples-all-in-one-with-options/1-smoke-test | job.batch/check-span created logger.go:42: 07:41:03 | examples-all-in-one-with-options/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:41:03 | examples-all-in-one-with-options | examples-all-in-one-with-options events from ns kuttl-test-enjoyed-mayfly: logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:39 +0000 UTC Normal Pod my-jaeger-6789bd78bf-2vmld Binding Scheduled Successfully assigned kuttl-test-enjoyed-mayfly/my-jaeger-6789bd78bf-2vmld to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:39 +0000 UTC Normal ReplicaSet.apps my-jaeger-6789bd78bf SuccessfulCreate Created pod: my-jaeger-6789bd78bf-2vmld replicaset-controller logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:39 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-6789bd78bf to 1 deployment-controller logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:40 +0000 UTC Normal Pod my-jaeger-6789bd78bf-2vmld AddedInterface Add eth0 [10.131.0.46/23] from ovn-kubernetes multus logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:40 +0000 UTC Normal Pod my-jaeger-6789bd78bf-2vmld.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:40 +0000 UTC Normal Pod my-jaeger-6789bd78bf-2vmld.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:40 +0000 UTC Normal Pod my-jaeger-6789bd78bf-2vmld.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:40 +0000 UTC Normal Pod my-jaeger-6789bd78bf-2vmld.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:40 +0000 UTC Normal Pod my-jaeger-6789bd78bf-2vmld.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:40 +0000 UTC Normal Pod my-jaeger-6789bd78bf-2vmld.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:44 +0000 UTC Normal Pod my-jaeger-6789bd78bf-2vmld.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:44 +0000 UTC Normal Pod my-jaeger-6789bd78bf-2vmld.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:44 +0000 UTC Normal ReplicaSet.apps my-jaeger-6789bd78bf SuccessfulDelete Deleted pod: my-jaeger-6789bd78bf-2vmld replicaset-controller logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:44 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-6789bd78bf to 0 from 1 deployment-controller logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:45 +0000 UTC Normal Pod my-jaeger-54b8cb888c-mvwtl Binding Scheduled Successfully assigned kuttl-test-enjoyed-mayfly/my-jaeger-54b8cb888c-mvwtl to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:45 +0000 UTC Normal ReplicaSet.apps my-jaeger-54b8cb888c SuccessfulCreate Created pod: my-jaeger-54b8cb888c-mvwtl replicaset-controller logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:45 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-54b8cb888c to 1 deployment-controller logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:46 +0000 UTC Normal Pod my-jaeger-54b8cb888c-mvwtl AddedInterface Add eth0 [10.131.0.47/23] from ovn-kubernetes multus logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:46 +0000 UTC Normal Pod my-jaeger-54b8cb888c-mvwtl.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:46 +0000 UTC Normal Pod my-jaeger-54b8cb888c-mvwtl.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:46 +0000 UTC Normal Pod my-jaeger-54b8cb888c-mvwtl.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:46 +0000 UTC Normal Pod my-jaeger-54b8cb888c-mvwtl.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:46 +0000 UTC Normal Pod my-jaeger-54b8cb888c-mvwtl.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:46 +0000 UTC Normal Pod my-jaeger-54b8cb888c-mvwtl.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:50 +0000 UTC Normal Pod check-span-hkwkx Binding Scheduled Successfully assigned kuttl-test-enjoyed-mayfly/check-span-hkwkx to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:50 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-hkwkx job-controller logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:50 +0000 UTC Normal Pod report-span-zpfj5 Binding Scheduled Successfully assigned kuttl-test-enjoyed-mayfly/report-span-zpfj5 to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:50 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-zpfj5 job-controller logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:51 +0000 UTC Normal Pod check-span-hkwkx AddedInterface Add eth0 [10.128.2.77/23] from ovn-kubernetes multus logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:51 +0000 UTC Normal Pod check-span-hkwkx.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:51 +0000 UTC Normal Pod report-span-zpfj5 AddedInterface Add eth0 [10.129.2.62/23] from ovn-kubernetes multus logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:51 +0000 UTC Normal Pod report-span-zpfj5.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:52 +0000 UTC Normal Pod check-span-hkwkx.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.313s (1.313s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:52 +0000 UTC Normal Pod check-span-hkwkx.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:52 +0000 UTC Normal Pod check-span-hkwkx.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:52 +0000 UTC Normal Pod report-span-zpfj5.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.39s (1.39s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:52 +0000 UTC Normal Pod report-span-zpfj5.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:40:52 +0000 UTC Normal Pod report-span-zpfj5.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:41:03 | examples-all-in-one-with-options | 2025-01-20 07:41:02 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:41:03 | examples-all-in-one-with-options | Deleting namespace: kuttl-test-enjoyed-mayfly === CONT kuttl/harness/examples-agent-with-priority-class logger.go:42: 07:41:16 | examples-agent-with-priority-class | Creating namespace: kuttl-test-loving-slug logger.go:42: 07:41:16 | examples-agent-with-priority-class/0-install | starting test step 0-install logger.go:42: 07:41:16 | examples-agent-with-priority-class/0-install | SecurityContextConstraints:/daemonset-with-hostport created logger.go:42: 07:41:16 | examples-agent-with-priority-class/0-install | ServiceAccount:kuttl-test-loving-slug/jaeger-agent-daemonset created logger.go:42: 07:41:16 | examples-agent-with-priority-class/0-install | test step completed 0-install logger.go:42: 07:41:16 | examples-agent-with-priority-class/1-add-policy | starting test step 1-add-policy logger.go:42: 07:41:16 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c oc adm policy --namespace $NAMESPACE add-scc-to-user daemonset-with-hostport -z jaeger-agent-daemonset] logger.go:42: 07:41:16 | examples-agent-with-priority-class/1-add-policy | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:daemonset-with-hostport added: "jaeger-agent-daemonset" logger.go:42: 07:41:16 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c sleep 5] logger.go:42: 07:41:21 | examples-agent-with-priority-class/1-add-policy | test step completed 1-add-policy logger.go:42: 07:41:21 | examples-agent-with-priority-class/2-install | starting test step 2-install logger.go:42: 07:41:21 | examples-agent-with-priority-class/2-install | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE agent-as-daemonset /dev/null] logger.go:42: 07:41:29 | examples-agent-with-priority-class/2-install | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:41:29 | examples-agent-with-priority-class/2-install | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:41:29 | examples-agent-with-priority-class/2-install | job.batch/report-span created logger.go:42: 07:41:29 | examples-agent-with-priority-class/2-install | job.batch/check-span created logger.go:42: 07:41:30 | examples-agent-with-priority-class/2-install | PriorityClass:/high-priority created logger.go:42: 07:41:30 | examples-agent-with-priority-class/2-install | Jaeger:kuttl-test-loving-slug/agent-as-daemonset updated logger.go:42: 07:41:43 | examples-agent-with-priority-class/2-install | test step completed 2-install logger.go:42: 07:41:43 | examples-agent-with-priority-class | examples-agent-with-priority-class events from ns kuttl-test-loving-slug: logger.go:42: 07:41:43 | examples-agent-with-priority-class | 2025-01-20 07:41:27 +0000 UTC Normal Pod agent-as-daemonset-db86d76c8-mjwdp Binding Scheduled Successfully assigned kuttl-test-loving-slug/agent-as-daemonset-db86d76c8-mjwdp to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:41:43 | examples-agent-with-priority-class | 2025-01-20 07:41:27 +0000 UTC Normal Pod agent-as-daemonset-db86d76c8-mjwdp AddedInterface Add eth0 [10.131.0.48/23] from ovn-kubernetes multus logger.go:42: 07:41:43 | examples-agent-with-priority-class | 2025-01-20 07:41:27 +0000 UTC Normal Pod agent-as-daemonset-db86d76c8-mjwdp.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:41:43 | examples-agent-with-priority-class | 2025-01-20 07:41:27 +0000 UTC Normal Pod agent-as-daemonset-db86d76c8-mjwdp.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:41:43 | examples-agent-with-priority-class | 2025-01-20 07:41:27 +0000 UTC Normal Pod agent-as-daemonset-db86d76c8-mjwdp.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:41:43 | examples-agent-with-priority-class | 2025-01-20 07:41:27 +0000 UTC Normal Pod agent-as-daemonset-db86d76c8-mjwdp.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:41:43 | examples-agent-with-priority-class | 2025-01-20 07:41:27 +0000 UTC Normal Pod agent-as-daemonset-db86d76c8-mjwdp.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:41:43 | examples-agent-with-priority-class | 2025-01-20 07:41:27 +0000 UTC Normal Pod agent-as-daemonset-db86d76c8-mjwdp.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:41:43 | examples-agent-with-priority-class | 2025-01-20 07:41:27 +0000 UTC Normal ReplicaSet.apps agent-as-daemonset-db86d76c8 SuccessfulCreate Created pod: agent-as-daemonset-db86d76c8-mjwdp replicaset-controller logger.go:42: 07:41:43 | examples-agent-with-priority-class | 2025-01-20 07:41:27 +0000 UTC Normal Deployment.apps agent-as-daemonset ScalingReplicaSet Scaled up replica set agent-as-daemonset-db86d76c8 to 1 deployment-controller logger.go:42: 07:41:43 | examples-agent-with-priority-class | 2025-01-20 07:41:29 +0000 UTC Normal Pod check-span-fmb47 Binding Scheduled Successfully assigned kuttl-test-loving-slug/check-span-fmb47 to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:41:43 | examples-agent-with-priority-class | 2025-01-20 07:41:29 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-fmb47 job-controller logger.go:42: 07:41:43 | examples-agent-with-priority-class | 2025-01-20 07:41:29 +0000 UTC Normal Pod report-span-x9jtc Binding Scheduled Successfully assigned kuttl-test-loving-slug/report-span-x9jtc to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:41:43 | examples-agent-with-priority-class | 2025-01-20 07:41:29 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-x9jtc job-controller logger.go:42: 07:41:43 | examples-agent-with-priority-class | 2025-01-20 07:41:30 +0000 UTC Normal Pod check-span-fmb47 AddedInterface Add eth0 [10.128.2.78/23] from ovn-kubernetes multus logger.go:42: 07:41:43 | examples-agent-with-priority-class | 2025-01-20 07:41:30 +0000 UTC Normal Pod check-span-fmb47.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:41:43 | examples-agent-with-priority-class | 2025-01-20 07:41:30 +0000 UTC Normal Pod report-span-x9jtc AddedInterface Add eth0 [10.129.2.63/23] from ovn-kubernetes multus logger.go:42: 07:41:43 | examples-agent-with-priority-class | 2025-01-20 07:41:30 +0000 UTC Normal Pod report-span-x9jtc.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:41:43 | examples-agent-with-priority-class | 2025-01-20 07:41:32 +0000 UTC Normal Pod check-span-fmb47.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.645s (1.645s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:41:43 | examples-agent-with-priority-class | 2025-01-20 07:41:32 +0000 UTC Normal Pod check-span-fmb47.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:41:43 | examples-agent-with-priority-class | 2025-01-20 07:41:32 +0000 UTC Normal Pod check-span-fmb47.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:41:43 | examples-agent-with-priority-class | 2025-01-20 07:41:32 +0000 UTC Normal Pod report-span-x9jtc.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.522s (1.522s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:41:43 | examples-agent-with-priority-class | 2025-01-20 07:41:32 +0000 UTC Normal Pod report-span-x9jtc.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:41:43 | examples-agent-with-priority-class | 2025-01-20 07:41:32 +0000 UTC Normal Pod report-span-x9jtc.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:41:43 | examples-agent-with-priority-class | 2025-01-20 07:41:33 +0000 UTC Warning DaemonSet.apps agent-as-daemonset-agent-daemonset FailedCreate Error creating: pods "agent-as-daemonset-agent-daemonset-" is forbidden: unable to validate against any security context constraint: [provider "anyuid": Forbidden: not usable by user or serviceaccount, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 5775: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 5778: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 6831: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 6832: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 14271: Host ports are not allowed to be used, provider "restricted": Forbidden: not usable by user or serviceaccount, provider "nonroot-v2": Forbidden: not usable by user or serviceaccount, provider "nonroot": Forbidden: not usable by user or serviceaccount, provider "hostmount-anyuid": Forbidden: not usable by user or serviceaccount, provider "elasticsearch-scc": Forbidden: not usable by user or serviceaccount, provider "machine-api-termination-handler": Forbidden: not usable by user or serviceaccount, provider "daemonset-with-hostport": Forbidden: not usable by user or serviceaccount, provider "hostnetwork-v2": Forbidden: not usable by user or serviceaccount, provider "hostnetwork": Forbidden: not usable by user or serviceaccount, provider "hostaccess": Forbidden: not usable by user or serviceaccount, provider "node-exporter": Forbidden: not usable by user or serviceaccount, provider "privileged": Forbidden: not usable by user or serviceaccount] daemonset-controller logger.go:42: 07:41:43 | examples-agent-with-priority-class | 2025-01-20 07:41:43 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:41:43 | examples-agent-with-priority-class | Deleting namespace: kuttl-test-loving-slug === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (534.30s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.00s) --- PASS: kuttl/harness/examples-simple-prod (71.38s) --- PASS: kuttl/harness/examples-with-sampling (58.07s) --- PASS: kuttl/harness/examples-with-cassandra (56.64s) --- PASS: kuttl/harness/examples-with-badger (33.72s) --- PASS: kuttl/harness/examples-simplest (36.96s) --- PASS: kuttl/harness/examples-simple-prod-with-volumes (70.79s) --- PASS: kuttl/harness/examples-business-application-injected-sidecar (38.65s) --- PASS: kuttl/harness/examples-service-types (57.28s) --- PASS: kuttl/harness/examples-openshift-with-htpasswd (23.96s) --- PASS: kuttl/harness/examples-all-in-one-with-options (40.95s) --- PASS: kuttl/harness/examples-agent-with-priority-class (39.74s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name examples --report --output /logs/artifacts/examples.xml ./artifacts/kuttl-report.xml time="2025-01-20T07:41:58Z" level=debug msg="Setting a new name for the test suites" time="2025-01-20T07:41:58Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-01-20T07:41:58Z" level=debug msg="normalizing test case names" time="2025-01-20T07:41:58Z" level=debug msg="examples/artifacts -> examples_artifacts" time="2025-01-20T07:41:58Z" level=debug msg="examples/examples-simple-prod -> examples_examples_simple_prod" time="2025-01-20T07:41:58Z" level=debug msg="examples/examples-with-sampling -> examples_examples_with_sampling" time="2025-01-20T07:41:58Z" level=debug msg="examples/examples-with-cassandra -> examples_examples_with_cassandra" time="2025-01-20T07:41:58Z" level=debug msg="examples/examples-with-badger -> examples_examples_with_badger" time="2025-01-20T07:41:58Z" level=debug msg="examples/examples-simplest -> examples_examples_simplest" time="2025-01-20T07:41:58Z" level=debug msg="examples/examples-simple-prod-with-volumes -> examples_examples_simple_prod_with_volumes" time="2025-01-20T07:41:58Z" level=debug msg="examples/examples-business-application-injected-sidecar -> examples_examples_business_application_injected_sidecar" time="2025-01-20T07:41:58Z" level=debug msg="examples/examples-service-types -> examples_examples_service_types" time="2025-01-20T07:41:58Z" level=debug msg="examples/examples-openshift-with-htpasswd -> examples_examples_openshift_with_htpasswd" time="2025-01-20T07:41:58Z" level=debug msg="examples/examples-all-in-one-with-options -> examples_examples_all_in_one_with_options" time="2025-01-20T07:41:58Z" level=debug msg="examples/examples-agent-with-priority-class -> examples_examples_agent_with_priority_class" +---------------------------------------------------------+--------+ | NAME | RESULT | +---------------------------------------------------------+--------+ | examples_artifacts | passed | | examples_examples_simple_prod | passed | | examples_examples_with_sampling | passed | | examples_examples_with_cassandra | passed | | examples_examples_with_badger | passed | | examples_examples_simplest | passed | | examples_examples_simple_prod_with_volumes | passed | | examples_examples_business_application_injected_sidecar | passed | | examples_examples_service_types | passed | | examples_examples_openshift_with_htpasswd | passed | | examples_examples_all_in_one_with_options | passed | | examples_examples_agent_with_priority_class | passed | +---------------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh generate false true + '[' 3 -ne 3 ']' + test_suite_name=generate + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/generate.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-generate make[2]: Entering directory '/tmp/jaeger-tests' test -s /tmp/jaeger-tests/bin/operator-sdk || curl -sLo /tmp/jaeger-tests/bin/operator-sdk https://github.com/operator-framework/operator-sdk/releases/download/v1.32.0/operator-sdk_`go env GOOS`_`go env GOARCH` ./hack/install/install-golangci-lint.sh Installing golangci-lint golangci-lint 1.55.2 is installed already ./hack/install/install-goimports.sh Installing goimports Try 0... go install golang.org/x/tools/cmd/goimports@v0.1.12 >>>> Formatting code... ./.ci/format.sh >>>> Building... ./hack/install/install-dependencies.sh Installing go dependencies Try 0... go mod download GOOS= GOARCH= CGO_ENABLED=0 GO111MODULE=on go build -ldflags "-X "github.com/jaegertracing/jaeger-operator/pkg/version".version="1.62.0" -X "github.com/jaegertracing/jaeger-operator/pkg/version".buildDate=2025-01-20T07:42:00Z -X "github.com/jaegertracing/jaeger-operator/pkg/version".defaultJaeger="1.62.0"" -o "bin/jaeger-operator" main.go JAEGER_VERSION="1.62.0" ./tests/e2e/generate/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-15-030107 True False 48m Cluster version is 4.18.0-0.nightly-2025-01-15-030107' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-15-030107 True False 48m Cluster version is 4.18.0-0.nightly-2025-01-15-030107' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/generate/render.sh ++ export SUITE_DIR=./tests/e2e/generate ++ SUITE_DIR=./tests/e2e/generate ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/generate ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + '[' true = true ']' + skip_test generate 'This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 2 -ne 2 ']' + test_name=generate + message='This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/generate/_build + '[' _build '!=' _build ']' + rm -rf generate + warning 'generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed\e[0m' WAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running generate E2E tests' Running generate E2E tests + cd tests/e2e/generate/_build + set +e + KUBECONFIG=/tmp/kubeconfig-3197734706 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 07:42:15 | artifacts | Creating namespace: kuttl-test-just-quagga logger.go:42: 07:42:15 | artifacts | artifacts events from ns kuttl-test-just-quagga: logger.go:42: 07:42:15 | artifacts | Deleting namespace: kuttl-test-just-quagga === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (6.08s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.90s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name generate --report --output /logs/artifacts/generate.xml ./artifacts/kuttl-report.xml time="2025-01-20T07:42:21Z" level=debug msg="Setting a new name for the test suites" time="2025-01-20T07:42:21Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-01-20T07:42:21Z" level=debug msg="normalizing test case names" time="2025-01-20T07:42:21Z" level=debug msg="generate/artifacts -> generate_artifacts" +--------------------+--------+ | NAME | RESULT | +--------------------+--------+ | generate_artifacts | passed | +--------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh sidecar false true + '[' 3 -ne 3 ']' + test_suite_name=sidecar + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/sidecar.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-sidecar make[2]: Entering directory '/tmp/jaeger-tests' ./tests/e2e/sidecar/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-15-030107 True False 48m Cluster version is 4.18.0-0.nightly-2025-01-15-030107' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-15-030107 True False 48m Cluster version is 4.18.0-0.nightly-2025-01-15-030107' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/sidecar/render.sh ++ export SUITE_DIR=./tests/e2e/sidecar ++ SUITE_DIR=./tests/e2e/sidecar ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/sidecar ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + jaeger_service_name=order + start_test sidecar-deployment + '[' 1 -ne 1 ']' + test_name=sidecar-deployment + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-deployment' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-deployment\e[0m' Rendering files for test sidecar-deployment + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build + '[' _build '!=' _build ']' + mkdir -p sidecar-deployment + cd sidecar-deployment + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-namespace + '[' 1 -ne 1 ']' + test_name=sidecar-namespace + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-namespace' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-namespace\e[0m' Rendering files for test sidecar-namespace + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-deployment + '[' sidecar-deployment '!=' _build ']' + cd .. + mkdir -p sidecar-namespace + cd sidecar-namespace + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-skip-webhook + '[' 1 -ne 1 ']' + test_name=sidecar-skip-webhook + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-skip-webhook' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-skip-webhook\e[0m' Rendering files for test sidecar-skip-webhook + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-namespace + '[' sidecar-namespace '!=' _build ']' + cd .. + mkdir -p sidecar-skip-webhook + cd sidecar-skip-webhook + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running sidecar E2E tests' Running sidecar E2E tests + cd tests/e2e/sidecar/_build + set +e + KUBECONFIG=/tmp/kubeconfig-3197734706 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 4 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/sidecar-deployment === PAUSE kuttl/harness/sidecar-deployment === RUN kuttl/harness/sidecar-namespace === PAUSE kuttl/harness/sidecar-namespace === RUN kuttl/harness/sidecar-skip-webhook === PAUSE kuttl/harness/sidecar-skip-webhook === CONT kuttl/harness/artifacts logger.go:42: 07:42:29 | artifacts | Creating namespace: kuttl-test-magnetic-elk logger.go:42: 07:42:29 | artifacts | artifacts events from ns kuttl-test-magnetic-elk: logger.go:42: 07:42:29 | artifacts | Deleting namespace: kuttl-test-magnetic-elk === CONT kuttl/harness/sidecar-namespace logger.go:42: 07:42:35 | sidecar-namespace | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:42:35 | sidecar-namespace | Creating namespace: kuttl-test-pleased-griffon logger.go:42: 07:42:35 | sidecar-namespace/0-install | starting test step 0-install logger.go:42: 07:42:35 | sidecar-namespace/0-install | Jaeger:kuttl-test-pleased-griffon/agent-as-sidecar created logger.go:42: 07:42:41 | sidecar-namespace/0-install | test step completed 0-install logger.go:42: 07:42:41 | sidecar-namespace/1-install | starting test step 1-install logger.go:42: 07:42:41 | sidecar-namespace/1-install | Deployment:kuttl-test-pleased-griffon/vertx-create-span-sidecar created logger.go:42: 07:42:44 | sidecar-namespace/1-install | test step completed 1-install logger.go:42: 07:42:44 | sidecar-namespace/2-enable-injection | starting test step 2-enable-injection logger.go:42: 07:42:44 | sidecar-namespace/2-enable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="true"] logger.go:42: 07:42:44 | sidecar-namespace/2-enable-injection | namespace/kuttl-test-pleased-griffon annotated logger.go:42: 07:42:48 | sidecar-namespace/2-enable-injection | test step completed 2-enable-injection logger.go:42: 07:42:48 | sidecar-namespace/3-find-service | starting test step 3-find-service logger.go:42: 07:42:48 | sidecar-namespace/3-find-service | Job:kuttl-test-pleased-griffon/00-find-service created logger.go:42: 07:43:03 | sidecar-namespace/3-find-service | test step completed 3-find-service logger.go:42: 07:43:03 | sidecar-namespace/4-other-instance | starting test step 4-other-instance logger.go:42: 07:43:03 | sidecar-namespace/4-other-instance | Jaeger:kuttl-test-pleased-griffon/agent-as-sidecar2 created logger.go:42: 07:43:12 | sidecar-namespace/4-other-instance | test step completed 4-other-instance logger.go:42: 07:43:12 | sidecar-namespace/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 07:43:13 | sidecar-namespace/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 07:43:13 | sidecar-namespace/6-find-service | starting test step 6-find-service logger.go:42: 07:43:13 | sidecar-namespace/6-find-service | Job:kuttl-test-pleased-griffon/01-find-service created logger.go:42: 07:43:33 | sidecar-namespace/6-find-service | test step completed 6-find-service logger.go:42: 07:43:33 | sidecar-namespace/7-disable-injection | starting test step 7-disable-injection logger.go:42: 07:43:33 | sidecar-namespace/7-disable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="false"] logger.go:42: 07:43:33 | sidecar-namespace/7-disable-injection | namespace/kuttl-test-pleased-griffon annotated logger.go:42: 07:43:35 | sidecar-namespace/7-disable-injection | test step completed 7-disable-injection logger.go:42: 07:43:35 | sidecar-namespace | sidecar-namespace events from ns kuttl-test-pleased-griffon: logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:39 +0000 UTC Normal Pod agent-as-sidecar-766f86bcc4-kh9s4 Binding Scheduled Successfully assigned kuttl-test-pleased-griffon/agent-as-sidecar-766f86bcc4-kh9s4 to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:39 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-766f86bcc4 SuccessfulCreate Created pod: agent-as-sidecar-766f86bcc4-kh9s4 replicaset-controller logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:39 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-766f86bcc4 to 1 deployment-controller logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:40 +0000 UTC Normal Pod agent-as-sidecar-766f86bcc4-kh9s4 AddedInterface Add eth0 [10.131.0.49/23] from ovn-kubernetes multus logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:40 +0000 UTC Normal Pod agent-as-sidecar-766f86bcc4-kh9s4.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:40 +0000 UTC Normal Pod agent-as-sidecar-766f86bcc4-kh9s4.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:40 +0000 UTC Normal Pod agent-as-sidecar-766f86bcc4-kh9s4.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:41 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-bzgjd Binding Scheduled Successfully assigned kuttl-test-pleased-griffon/vertx-create-span-sidecar-797645c8fc-bzgjd to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:41 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-797645c8fc SuccessfulCreate Created pod: vertx-create-span-sidecar-797645c8fc-bzgjd replicaset-controller logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:41 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-797645c8fc to 1 deployment-controller logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:42 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-bzgjd AddedInterface Add eth0 [10.129.2.64/23] from ovn-kubernetes multus logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:42 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-bzgjd.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:42 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-bzgjd.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:42 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-bzgjd.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:44 +0000 UTC Normal Pod vertx-create-span-sidecar-97f65cbc7-cscgc Binding Scheduled Successfully assigned kuttl-test-pleased-griffon/vertx-create-span-sidecar-97f65cbc7-cscgc to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:44 +0000 UTC Normal Pod vertx-create-span-sidecar-97f65cbc7-cscgc AddedInterface Add eth0 [10.128.2.79/23] from ovn-kubernetes multus logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:44 +0000 UTC Normal Pod vertx-create-span-sidecar-97f65cbc7-cscgc.spec.containers{vertx-create-span-sidecar} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:44 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-97f65cbc7 SuccessfulCreate Created pod: vertx-create-span-sidecar-97f65cbc7-cscgc replicaset-controller logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:44 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-97f65cbc7 to 1 deployment-controller logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:47 +0000 UTC Normal Pod vertx-create-span-sidecar-97f65cbc7-cscgc.spec.containers{vertx-create-span-sidecar} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 2.987s (2.987s including waiting). Image size: 282912835 bytes. kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:47 +0000 UTC Normal Pod vertx-create-span-sidecar-97f65cbc7-cscgc.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:47 +0000 UTC Normal Pod vertx-create-span-sidecar-97f65cbc7-cscgc.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:47 +0000 UTC Normal Pod vertx-create-span-sidecar-97f65cbc7-cscgc.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:47 +0000 UTC Normal Pod vertx-create-span-sidecar-97f65cbc7-cscgc.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:47 +0000 UTC Normal Pod vertx-create-span-sidecar-97f65cbc7-cscgc.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:48 +0000 UTC Normal Pod 00-find-service-9mg5q Binding Scheduled Successfully assigned kuttl-test-pleased-griffon/00-find-service-9mg5q to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:48 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-9mg5q job-controller logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:49 +0000 UTC Normal Pod 00-find-service-9mg5q AddedInterface Add eth0 [10.131.0.50/23] from ovn-kubernetes multus logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:49 +0000 UTC Normal Pod 00-find-service-9mg5q.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:50 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-bzgjd.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.64:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:50 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-bzgjd.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.64:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:51 +0000 UTC Normal Pod 00-find-service-9mg5q.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 2.901s (2.901s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:52 +0000 UTC Normal Pod 00-find-service-9mg5q.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:52 +0000 UTC Normal Pod 00-find-service-9mg5q.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:52 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-bzgjd.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:52 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-bzgjd.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.64:8080/": read tcp 10.129.2.2:58982->10.129.2.64:8080: read: connection reset by peer kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:52 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-bzgjd.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.64:8080/": dial tcp 10.129.2.64:8080: connect: connection refused kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:55 +0000 UTC Warning Pod vertx-create-span-sidecar-97f65cbc7-cscgc.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.128.2.79:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:55 +0000 UTC Warning Pod vertx-create-span-sidecar-97f65cbc7-cscgc.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.79:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:57 +0000 UTC Normal Pod vertx-create-span-sidecar-97f65cbc7-cscgc.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:58 +0000 UTC Warning Pod vertx-create-span-sidecar-97f65cbc7-cscgc.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.79:8080/": read tcp 10.128.2.2:37924->10.128.2.79:8080: read: connection reset by peer kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:58 +0000 UTC Warning Pod vertx-create-span-sidecar-97f65cbc7-cscgc.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.79:8080/": dial tcp 10.128.2.79:8080: connect: connection refused kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:42:58 +0000 UTC Normal Pod vertx-create-span-sidecar-97f65cbc7-cscgc.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:02 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:02 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-bzgjd.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.64:8080/": read tcp 10.129.2.2:34846->10.129.2.64:8080: read: connection reset by peer kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:10 +0000 UTC Normal Pod agent-as-sidecar2-74499fd8-jksq2 Binding Scheduled Successfully assigned kuttl-test-pleased-griffon/agent-as-sidecar2-74499fd8-jksq2 to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:10 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-74499fd8 SuccessfulCreate Created pod: agent-as-sidecar2-74499fd8-jksq2 replicaset-controller logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:10 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-74499fd8 to 1 deployment-controller logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:11 +0000 UTC Normal Pod agent-as-sidecar2-74499fd8-jksq2 AddedInterface Add eth0 [10.131.0.51/23] from ovn-kubernetes multus logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:11 +0000 UTC Normal Pod agent-as-sidecar2-74499fd8-jksq2.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:11 +0000 UTC Normal Pod agent-as-sidecar2-74499fd8-jksq2.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:11 +0000 UTC Normal Pod agent-as-sidecar2-74499fd8-jksq2.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:13 +0000 UTC Normal Pod 01-find-service-s7qk7 Binding Scheduled Successfully assigned kuttl-test-pleased-griffon/01-find-service-s7qk7 to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:13 +0000 UTC Normal Pod 01-find-service-s7qk7 AddedInterface Add eth0 [10.129.2.65/23] from ovn-kubernetes multus logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:13 +0000 UTC Normal Pod 01-find-service-s7qk7.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:13 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-s7qk7 job-controller logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:13 +0000 UTC Normal Pod agent-as-sidecar-766f86bcc4-kh9s4.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:14 +0000 UTC Normal Pod 01-find-service-s7qk7.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 485ms (485ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:14 +0000 UTC Normal Pod 01-find-service-s7qk7.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:14 +0000 UTC Normal Pod 01-find-service-s7qk7.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:15 +0000 UTC Normal Pod vertx-create-span-sidecar-6b9459cb5f-8gchl Binding Scheduled Successfully assigned kuttl-test-pleased-griffon/vertx-create-span-sidecar-6b9459cb5f-8gchl to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:15 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-6b9459cb5f SuccessfulCreate Created pod: vertx-create-span-sidecar-6b9459cb5f-8gchl replicaset-controller logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:15 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-797645c8fc SuccessfulDelete Deleted pod: vertx-create-span-sidecar-797645c8fc-bzgjd replicaset-controller logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:15 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-797645c8fc to 0 from 1 deployment-controller logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:15 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-6b9459cb5f to 1 from 0 deployment-controller logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:16 +0000 UTC Normal Pod vertx-create-span-sidecar-6b9459cb5f-8gchl AddedInterface Add eth0 [10.131.0.52/23] from ovn-kubernetes multus logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:16 +0000 UTC Normal Pod vertx-create-span-sidecar-6b9459cb5f-8gchl.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:16 +0000 UTC Normal Pod vertx-create-span-sidecar-6b9459cb5f-8gchl.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:16 +0000 UTC Normal Pod vertx-create-span-sidecar-6b9459cb5f-8gchl.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:16 +0000 UTC Normal Pod vertx-create-span-sidecar-6b9459cb5f-8gchl.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:17 +0000 UTC Normal Pod vertx-create-span-sidecar-6b9459cb5f-8gchl.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" in 1.667s (1.667s including waiting). Image size: 115697219 bytes. kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:17 +0000 UTC Normal Pod vertx-create-span-sidecar-6b9459cb5f-8gchl.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:17 +0000 UTC Normal Pod vertx-create-span-sidecar-6b9459cb5f-8gchl.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:24 +0000 UTC Warning Pod vertx-create-span-sidecar-6b9459cb5f-8gchl.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.52:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:24 +0000 UTC Warning Pod vertx-create-span-sidecar-6b9459cb5f-8gchl.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.131.0.52:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:27 +0000 UTC Normal Pod vertx-create-span-sidecar-6b9459cb5f-8gchl.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:27 +0000 UTC Warning Pod vertx-create-span-sidecar-6b9459cb5f-8gchl.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.52:8080/": read tcp 10.131.0.2:60834->10.131.0.52:8080: read: connection reset by peer kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:27 +0000 UTC Warning Pod vertx-create-span-sidecar-6b9459cb5f-8gchl.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.52:8080/": dial tcp 10.131.0.52:8080: connect: connection refused kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:33 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:33 +0000 UTC Normal Pod vertx-create-span-sidecar-886d5c6f6-72nfd Binding Scheduled Successfully assigned kuttl-test-pleased-griffon/vertx-create-span-sidecar-886d5c6f6-72nfd to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:33 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-886d5c6f6 SuccessfulCreate Created pod: vertx-create-span-sidecar-886d5c6f6-72nfd replicaset-controller logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:33 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-97f65cbc7 SuccessfulDelete Deleted pod: vertx-create-span-sidecar-97f65cbc7-cscgc replicaset-controller logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:33 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-97f65cbc7 to 0 from 1 deployment-controller logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:33 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-886d5c6f6 to 1 from 0 deployment-controller logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:34 +0000 UTC Normal Pod vertx-create-span-sidecar-886d5c6f6-72nfd AddedInterface Add eth0 [10.129.2.66/23] from ovn-kubernetes multus logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:34 +0000 UTC Normal Pod vertx-create-span-sidecar-886d5c6f6-72nfd.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:34 +0000 UTC Normal Pod vertx-create-span-sidecar-886d5c6f6-72nfd.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:43:35 | sidecar-namespace | 2025-01-20 07:43:34 +0000 UTC Normal Pod vertx-create-span-sidecar-886d5c6f6-72nfd.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:43:36 | sidecar-namespace | Deleting namespace: kuttl-test-pleased-griffon === CONT kuttl/harness/sidecar-skip-webhook logger.go:42: 07:43:42 | sidecar-skip-webhook | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:43:42 | sidecar-skip-webhook | Creating namespace: kuttl-test-promoted-pheasant logger.go:42: 07:43:42 | sidecar-skip-webhook/0-install | starting test step 0-install logger.go:42: 07:43:42 | sidecar-skip-webhook/0-install | Jaeger:kuttl-test-promoted-pheasant/agent-as-sidecar created logger.go:42: 07:43:50 | sidecar-skip-webhook/0-install | test step completed 0-install logger.go:42: 07:43:50 | sidecar-skip-webhook/1-install | starting test step 1-install logger.go:42: 07:43:50 | sidecar-skip-webhook/1-install | Deployment:kuttl-test-promoted-pheasant/vertx-create-span-sidecar created logger.go:42: 07:43:52 | sidecar-skip-webhook/1-install | test step completed 1-install logger.go:42: 07:43:52 | sidecar-skip-webhook/2-add-anotation-and-label | starting test step 2-add-anotation-and-label logger.go:42: 07:43:52 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name=jaeger-operator --namespace kuttl-test-promoted-pheasant] logger.go:42: 07:43:52 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar labeled logger.go:42: 07:43:52 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-promoted-pheasant] logger.go:42: 07:43:52 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 07:43:52 | sidecar-skip-webhook/2-add-anotation-and-label | test step completed 2-add-anotation-and-label logger.go:42: 07:43:52 | sidecar-skip-webhook/3-remove-label | starting test step 3-remove-label logger.go:42: 07:43:52 | sidecar-skip-webhook/3-remove-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name- --namespace kuttl-test-promoted-pheasant] logger.go:42: 07:43:52 | sidecar-skip-webhook/3-remove-label | deployment.apps/vertx-create-span-sidecar unlabeled logger.go:42: 07:43:54 | sidecar-skip-webhook/3-remove-label | test step completed 3-remove-label logger.go:42: 07:43:54 | sidecar-skip-webhook | sidecar-skip-webhook events from ns kuttl-test-promoted-pheasant: logger.go:42: 07:43:54 | sidecar-skip-webhook | 2025-01-20 07:43:46 +0000 UTC Normal Pod agent-as-sidecar-7c84c6bb5f-n2kxj Binding Scheduled Successfully assigned kuttl-test-promoted-pheasant/agent-as-sidecar-7c84c6bb5f-n2kxj to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:43:54 | sidecar-skip-webhook | 2025-01-20 07:43:46 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-7c84c6bb5f SuccessfulCreate Created pod: agent-as-sidecar-7c84c6bb5f-n2kxj replicaset-controller logger.go:42: 07:43:54 | sidecar-skip-webhook | 2025-01-20 07:43:46 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-7c84c6bb5f to 1 deployment-controller logger.go:42: 07:43:54 | sidecar-skip-webhook | 2025-01-20 07:43:47 +0000 UTC Warning Pod agent-as-sidecar-7c84c6bb5f-n2kxj FailedMount MountVolume.SetUp failed for volume "agent-as-sidecar-collector-tls-config-volume" : secret "agent-as-sidecar-collector-headless-tls" not found kubelet logger.go:42: 07:43:54 | sidecar-skip-webhook | 2025-01-20 07:43:48 +0000 UTC Normal Pod agent-as-sidecar-7c84c6bb5f-n2kxj AddedInterface Add eth0 [10.131.0.53/23] from ovn-kubernetes multus logger.go:42: 07:43:54 | sidecar-skip-webhook | 2025-01-20 07:43:48 +0000 UTC Normal Pod agent-as-sidecar-7c84c6bb5f-n2kxj.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:43:54 | sidecar-skip-webhook | 2025-01-20 07:43:48 +0000 UTC Normal Pod agent-as-sidecar-7c84c6bb5f-n2kxj.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:43:54 | sidecar-skip-webhook | 2025-01-20 07:43:48 +0000 UTC Normal Pod agent-as-sidecar-7c84c6bb5f-n2kxj.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:43:54 | sidecar-skip-webhook | 2025-01-20 07:43:50 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-fkztf Binding Scheduled Successfully assigned kuttl-test-promoted-pheasant/vertx-create-span-sidecar-797645c8fc-fkztf to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:43:54 | sidecar-skip-webhook | 2025-01-20 07:43:50 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-fkztf AddedInterface Add eth0 [10.129.2.67/23] from ovn-kubernetes multus logger.go:42: 07:43:54 | sidecar-skip-webhook | 2025-01-20 07:43:50 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-fkztf.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:43:54 | sidecar-skip-webhook | 2025-01-20 07:43:50 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-fkztf.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:43:54 | sidecar-skip-webhook | 2025-01-20 07:43:50 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-fkztf.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:43:54 | sidecar-skip-webhook | 2025-01-20 07:43:50 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-797645c8fc SuccessfulCreate Created pod: vertx-create-span-sidecar-797645c8fc-fkztf replicaset-controller logger.go:42: 07:43:54 | sidecar-skip-webhook | 2025-01-20 07:43:50 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-797645c8fc to 1 deployment-controller logger.go:42: 07:43:54 | sidecar-skip-webhook | 2025-01-20 07:43:52 +0000 UTC Normal Pod vertx-create-span-sidecar-56f5fb7768-4r5jh Binding Scheduled Successfully assigned kuttl-test-promoted-pheasant/vertx-create-span-sidecar-56f5fb7768-4r5jh to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:43:54 | sidecar-skip-webhook | 2025-01-20 07:43:52 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-56f5fb7768 SuccessfulCreate Created pod: vertx-create-span-sidecar-56f5fb7768-4r5jh replicaset-controller logger.go:42: 07:43:54 | sidecar-skip-webhook | 2025-01-20 07:43:52 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-56f5fb7768 to 1 deployment-controller logger.go:42: 07:43:54 | sidecar-skip-webhook | 2025-01-20 07:43:53 +0000 UTC Normal Pod vertx-create-span-sidecar-56f5fb7768-4r5jh AddedInterface Add eth0 [10.131.0.54/23] from ovn-kubernetes multus logger.go:42: 07:43:54 | sidecar-skip-webhook | 2025-01-20 07:43:53 +0000 UTC Normal Pod vertx-create-span-sidecar-56f5fb7768-4r5jh.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:43:54 | sidecar-skip-webhook | 2025-01-20 07:43:53 +0000 UTC Normal Pod vertx-create-span-sidecar-56f5fb7768-4r5jh.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:43:54 | sidecar-skip-webhook | 2025-01-20 07:43:53 +0000 UTC Normal Pod vertx-create-span-sidecar-56f5fb7768-4r5jh.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:43:54 | sidecar-skip-webhook | 2025-01-20 07:43:53 +0000 UTC Normal Pod vertx-create-span-sidecar-56f5fb7768-4r5jh.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:43:54 | sidecar-skip-webhook | 2025-01-20 07:43:53 +0000 UTC Normal Pod vertx-create-span-sidecar-56f5fb7768-4r5jh.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:43:54 | sidecar-skip-webhook | 2025-01-20 07:43:53 +0000 UTC Normal Pod vertx-create-span-sidecar-56f5fb7768-4r5jh.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:43:54 | sidecar-skip-webhook | Deleting namespace: kuttl-test-promoted-pheasant === CONT kuttl/harness/sidecar-deployment logger.go:42: 07:44:01 | sidecar-deployment | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:44:01 | sidecar-deployment | Creating namespace: kuttl-test-meet-anchovy logger.go:42: 07:44:01 | sidecar-deployment/0-install | starting test step 0-install logger.go:42: 07:44:01 | sidecar-deployment/0-install | Jaeger:kuttl-test-meet-anchovy/agent-as-sidecar created logger.go:42: 07:44:07 | sidecar-deployment/0-install | test step completed 0-install logger.go:42: 07:44:07 | sidecar-deployment/1-install | starting test step 1-install logger.go:42: 07:44:07 | sidecar-deployment/1-install | Deployment:kuttl-test-meet-anchovy/vertx-create-span-sidecar created logger.go:42: 07:44:08 | sidecar-deployment/1-install | test step completed 1-install logger.go:42: 07:44:08 | sidecar-deployment/2-enable-injection | starting test step 2-enable-injection logger.go:42: 07:44:08 | sidecar-deployment/2-enable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-meet-anchovy] logger.go:42: 07:44:08 | sidecar-deployment/2-enable-injection | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 07:44:10 | sidecar-deployment/2-enable-injection | test step completed 2-enable-injection logger.go:42: 07:44:10 | sidecar-deployment/3-find-service | starting test step 3-find-service logger.go:42: 07:44:11 | sidecar-deployment/3-find-service | Job:kuttl-test-meet-anchovy/00-find-service created logger.go:42: 07:44:23 | sidecar-deployment/3-find-service | test step completed 3-find-service logger.go:42: 07:44:23 | sidecar-deployment/4-other-instance | starting test step 4-other-instance logger.go:42: 07:44:23 | sidecar-deployment/4-other-instance | Jaeger:kuttl-test-meet-anchovy/agent-as-sidecar2 created logger.go:42: 07:44:29 | sidecar-deployment/4-other-instance | test step completed 4-other-instance logger.go:42: 07:44:29 | sidecar-deployment/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 07:44:29 | sidecar-deployment/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 07:44:29 | sidecar-deployment/6-find-service | starting test step 6-find-service logger.go:42: 07:44:29 | sidecar-deployment/6-find-service | Job:kuttl-test-meet-anchovy/01-find-service created logger.go:42: 07:44:43 | sidecar-deployment/6-find-service | test step completed 6-find-service logger.go:42: 07:44:43 | sidecar-deployment/7-disable-injection | starting test step 7-disable-injection logger.go:42: 07:44:43 | sidecar-deployment/7-disable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=false --namespace kuttl-test-meet-anchovy] logger.go:42: 07:44:44 | sidecar-deployment/7-disable-injection | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 07:44:46 | sidecar-deployment/7-disable-injection | test step completed 7-disable-injection logger.go:42: 07:44:46 | sidecar-deployment | sidecar-deployment events from ns kuttl-test-meet-anchovy: logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:05 +0000 UTC Normal Pod agent-as-sidecar-7d69b848f6-tb8xx Binding Scheduled Successfully assigned kuttl-test-meet-anchovy/agent-as-sidecar-7d69b848f6-tb8xx to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:05 +0000 UTC Normal Pod agent-as-sidecar-7d69b848f6-tb8xx AddedInterface Add eth0 [10.131.0.55/23] from ovn-kubernetes multus logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:05 +0000 UTC Normal Pod agent-as-sidecar-7d69b848f6-tb8xx.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:05 +0000 UTC Normal Pod agent-as-sidecar-7d69b848f6-tb8xx.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:05 +0000 UTC Normal Pod agent-as-sidecar-7d69b848f6-tb8xx.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:05 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-7d69b848f6 SuccessfulCreate Created pod: agent-as-sidecar-7d69b848f6-tb8xx replicaset-controller logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:05 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-7d69b848f6 to 1 deployment-controller logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:07 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-jfsv5 Binding Scheduled Successfully assigned kuttl-test-meet-anchovy/vertx-create-span-sidecar-797645c8fc-jfsv5 to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:07 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-797645c8fc SuccessfulCreate Created pod: vertx-create-span-sidecar-797645c8fc-jfsv5 replicaset-controller logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:07 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-797645c8fc to 1 deployment-controller logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:08 +0000 UTC Normal Pod vertx-create-span-sidecar-54f67b8555-zmfgz Binding Scheduled Successfully assigned kuttl-test-meet-anchovy/vertx-create-span-sidecar-54f67b8555-zmfgz to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:08 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-54f67b8555 SuccessfulCreate Created pod: vertx-create-span-sidecar-54f67b8555-zmfgz replicaset-controller logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:08 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-jfsv5 AddedInterface Add eth0 [10.129.2.68/23] from ovn-kubernetes multus logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:08 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-jfsv5.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:08 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-jfsv5.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:08 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-jfsv5.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:08 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-54f67b8555 to 1 deployment-controller logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:09 +0000 UTC Normal Pod vertx-create-span-sidecar-54f67b8555-zmfgz AddedInterface Add eth0 [10.128.2.80/23] from ovn-kubernetes multus logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:09 +0000 UTC Normal Pod vertx-create-span-sidecar-54f67b8555-zmfgz.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:09 +0000 UTC Normal Pod vertx-create-span-sidecar-54f67b8555-zmfgz.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:09 +0000 UTC Normal Pod vertx-create-span-sidecar-54f67b8555-zmfgz.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:09 +0000 UTC Normal Pod vertx-create-span-sidecar-54f67b8555-zmfgz.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:09 +0000 UTC Normal Pod vertx-create-span-sidecar-54f67b8555-zmfgz.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:09 +0000 UTC Normal Pod vertx-create-span-sidecar-54f67b8555-zmfgz.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:11 +0000 UTC Normal Pod 00-find-service-vcc86 Binding Scheduled Successfully assigned kuttl-test-meet-anchovy/00-find-service-vcc86 to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:11 +0000 UTC Normal Pod 00-find-service-vcc86 AddedInterface Add eth0 [10.131.0.56/23] from ovn-kubernetes multus logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:11 +0000 UTC Normal Pod 00-find-service-vcc86.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:11 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-vcc86 job-controller logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:12 +0000 UTC Normal Pod 00-find-service-vcc86.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 510ms (510ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:12 +0000 UTC Normal Pod 00-find-service-vcc86.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:12 +0000 UTC Normal Pod 00-find-service-vcc86.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:16 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-jfsv5.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.68:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:17 +0000 UTC Warning Pod vertx-create-span-sidecar-54f67b8555-zmfgz.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.80:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:17 +0000 UTC Warning Pod vertx-create-span-sidecar-54f67b8555-zmfgz.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.128.2.80:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:17 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-jfsv5.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.68:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:19 +0000 UTC Normal Pod vertx-create-span-sidecar-54f67b8555-zmfgz.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:19 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-jfsv5.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:19 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-jfsv5.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.68:8080/": read tcp 10.129.2.2:36542->10.129.2.68:8080: read: connection reset by peer kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:19 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-jfsv5.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.68:8080/": dial tcp 10.129.2.68:8080: connect: connection refused kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:20 +0000 UTC Warning Pod vertx-create-span-sidecar-54f67b8555-zmfgz.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.80:8080/": read tcp 10.128.2.2:58200->10.128.2.80:8080: read: connection reset by peer kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:20 +0000 UTC Warning Pod vertx-create-span-sidecar-54f67b8555-zmfgz.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.80:8080/": dial tcp 10.128.2.80:8080: connect: connection refused kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:23 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:27 +0000 UTC Normal Pod agent-as-sidecar2-5dfc6fc4-7x4md Binding Scheduled Successfully assigned kuttl-test-meet-anchovy/agent-as-sidecar2-5dfc6fc4-7x4md to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:27 +0000 UTC Normal Pod agent-as-sidecar2-5dfc6fc4-7x4md AddedInterface Add eth0 [10.131.0.57/23] from ovn-kubernetes multus logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:27 +0000 UTC Normal Pod agent-as-sidecar2-5dfc6fc4-7x4md.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:27 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-5dfc6fc4 SuccessfulCreate Created pod: agent-as-sidecar2-5dfc6fc4-7x4md replicaset-controller logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:27 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-5dfc6fc4 to 1 deployment-controller logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:28 +0000 UTC Normal Pod agent-as-sidecar2-5dfc6fc4-7x4md.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:28 +0000 UTC Normal Pod agent-as-sidecar2-5dfc6fc4-7x4md.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:29 +0000 UTC Normal Pod 01-find-service-qmq27 Binding Scheduled Successfully assigned kuttl-test-meet-anchovy/01-find-service-qmq27 to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:29 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-qmq27 job-controller logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:29 +0000 UTC Normal Pod agent-as-sidecar-7d69b848f6-tb8xx.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:30 +0000 UTC Normal Pod 01-find-service-qmq27 AddedInterface Add eth0 [10.129.2.69/23] from ovn-kubernetes multus logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:30 +0000 UTC Normal Pod 01-find-service-qmq27.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:30 +0000 UTC Normal Pod vertx-create-span-sidecar-76674cd69-xt58j Binding Scheduled Successfully assigned kuttl-test-meet-anchovy/vertx-create-span-sidecar-76674cd69-xt58j to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:30 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-76674cd69 SuccessfulCreate Created pod: vertx-create-span-sidecar-76674cd69-xt58j replicaset-controller logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:30 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-jfsv5.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.68:8080/": read tcp 10.129.2.2:45808->10.129.2.68:8080: read: connection reset by peer kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:30 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-797645c8fc SuccessfulDelete Deleted pod: vertx-create-span-sidecar-797645c8fc-jfsv5 replicaset-controller logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:30 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-797645c8fc to 0 from 1 deployment-controller logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:30 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-76674cd69 to 1 from 0 deployment-controller logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:31 +0000 UTC Normal Pod vertx-create-span-sidecar-76674cd69-xt58j AddedInterface Add eth0 [10.131.0.58/23] from ovn-kubernetes multus logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:31 +0000 UTC Normal Pod vertx-create-span-sidecar-76674cd69-xt58j.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:31 +0000 UTC Normal Pod vertx-create-span-sidecar-76674cd69-xt58j.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:31 +0000 UTC Normal Pod vertx-create-span-sidecar-76674cd69-xt58j.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:31 +0000 UTC Normal Pod vertx-create-span-sidecar-76674cd69-xt58j.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:31 +0000 UTC Normal Pod vertx-create-span-sidecar-76674cd69-xt58j.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:31 +0000 UTC Normal Pod vertx-create-span-sidecar-76674cd69-xt58j.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:33 +0000 UTC Normal Pod 01-find-service-qmq27.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 3.02s (3.02s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:33 +0000 UTC Normal Pod 01-find-service-qmq27.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:33 +0000 UTC Normal Pod 01-find-service-qmq27.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:39 +0000 UTC Warning Pod vertx-create-span-sidecar-76674cd69-xt58j.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.131.0.58:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:39 +0000 UTC Warning Pod vertx-create-span-sidecar-76674cd69-xt58j.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.58:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:42 +0000 UTC Normal Pod vertx-create-span-sidecar-76674cd69-xt58j.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:42 +0000 UTC Warning Pod vertx-create-span-sidecar-76674cd69-xt58j.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.58:8080/": read tcp 10.131.0.2:38334->10.131.0.58:8080: read: connection reset by peer kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:42 +0000 UTC Warning Pod vertx-create-span-sidecar-76674cd69-xt58j.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.58:8080/": dial tcp 10.131.0.58:8080: connect: connection refused kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:43 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:44 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-54f67b8555 SuccessfulDelete Deleted pod: vertx-create-span-sidecar-54f67b8555-zmfgz replicaset-controller logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:44 +0000 UTC Normal Pod vertx-create-span-sidecar-787676b7b6-kmjgl Binding Scheduled Successfully assigned kuttl-test-meet-anchovy/vertx-create-span-sidecar-787676b7b6-kmjgl to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:44 +0000 UTC Normal Pod vertx-create-span-sidecar-787676b7b6-kmjgl AddedInterface Add eth0 [10.129.2.70/23] from ovn-kubernetes multus logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:44 +0000 UTC Normal Pod vertx-create-span-sidecar-787676b7b6-kmjgl.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:44 +0000 UTC Normal Pod vertx-create-span-sidecar-787676b7b6-kmjgl.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:44 +0000 UTC Normal Pod vertx-create-span-sidecar-787676b7b6-kmjgl.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:44 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-787676b7b6 SuccessfulCreate Created pod: vertx-create-span-sidecar-787676b7b6-kmjgl replicaset-controller logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:44 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-54f67b8555 to 0 from 1 deployment-controller logger.go:42: 07:44:46 | sidecar-deployment | 2025-01-20 07:44:44 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-787676b7b6 to 1 from 0 deployment-controller logger.go:42: 07:44:46 | sidecar-deployment | Deleting namespace: kuttl-test-meet-anchovy === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (143.55s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.93s) --- PASS: kuttl/harness/sidecar-namespace (67.22s) --- PASS: kuttl/harness/sidecar-skip-webhook (18.46s) --- PASS: kuttl/harness/sidecar-deployment (51.78s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name sidecar --report --output /logs/artifacts/sidecar.xml ./artifacts/kuttl-report.xml time="2025-01-20T07:44:53Z" level=debug msg="Setting a new name for the test suites" time="2025-01-20T07:44:53Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-01-20T07:44:53Z" level=debug msg="normalizing test case names" time="2025-01-20T07:44:53Z" level=debug msg="sidecar/artifacts -> sidecar_artifacts" time="2025-01-20T07:44:53Z" level=debug msg="sidecar/sidecar-namespace -> sidecar_sidecar_namespace" time="2025-01-20T07:44:53Z" level=debug msg="sidecar/sidecar-skip-webhook -> sidecar_sidecar_skip_webhook" time="2025-01-20T07:44:53Z" level=debug msg="sidecar/sidecar-deployment -> sidecar_sidecar_deployment" +------------------------------+--------+ | NAME | RESULT | +------------------------------+--------+ | sidecar_artifacts | passed | | sidecar_sidecar_namespace | passed | | sidecar_sidecar_skip_webhook | passed | | sidecar_sidecar_deployment | passed | +------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh streaming false true + '[' 3 -ne 3 ']' + test_suite_name=streaming + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/streaming.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-streaming make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ SKIP_ES_EXTERNAL=true \ ./tests/e2e/streaming/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-15-030107 True False 50m Cluster version is 4.18.0-0.nightly-2025-01-15-030107' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-15-030107 True False 50m Cluster version is 4.18.0-0.nightly-2025-01-15-030107' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/streaming/render.sh ++ export SUITE_DIR=./tests/e2e/streaming ++ SUITE_DIR=./tests/e2e/streaming ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/streaming ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + '[' false = true ']' + start_test streaming-simple + '[' 1 -ne 1 ']' + test_name=streaming-simple + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-simple' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-simple\e[0m' Rendering files for test streaming-simple + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build + '[' _build '!=' _build ']' + mkdir -p streaming-simple + cd streaming-simple + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/streaming-jaeger-assert.yaml.template -o ./04-assert.yaml + render_smoke_test simple-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=simple-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + export JAEGER_NAME=simple-streaming + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-tls + '[' 1 -ne 1 ']' + test_name=streaming-with-tls + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-tls' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-tls\e[0m' Rendering files for test streaming-with-tls + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-simple + '[' streaming-simple '!=' _build ']' + cd .. + mkdir -p streaming-with-tls + cd streaming-with-tls + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + render_smoke_test tls-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=tls-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + export JAEGER_NAME=tls-streaming + JAEGER_NAME=tls-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-autoprovisioning-autoscale + '[' 1 -ne 1 ']' + test_name=streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-autoprovisioning-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-autoprovisioning-autoscale\e[0m' Rendering files for test streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-with-tls + '[' streaming-with-tls '!=' _build ']' + cd .. + mkdir -p streaming-with-autoprovisioning-autoscale + cd streaming-with-autoprovisioning-autoscale + '[' true = true ']' + rm ./00-install.yaml ./00-assert.yaml + render_install_elasticsearch upstream 01 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=01 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./01-assert.yaml + jaeger_name=auto-provisioned + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="20Mi"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="500m"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.autoscale=true ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.minReplicas=1 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.maxReplicas=2 ./02-install.yaml + render_assert_kafka true auto-provisioned 03 + '[' 3 -ne 3 ']' + autoprovisioned=true + cluster_name=auto-provisioned + test_step=03 + '[' true = true ']' + is_kafka_minimal_enabled + namespaces=(observability openshift-operators openshift-distributed-tracing) + for i in "${namespaces[@]}" ++ kubectl get pods -n observability -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-operators -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-distributed-tracing -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled=true + '[' true == true ']' + return 0 + replicas=1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./03-assert.yaml ++ expr 03 + 1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./04-assert.yaml ++ expr 03 + 2 + CLUSTER_NAME=auto-provisioned + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./05-assert.yaml + version_lt 1.30 1.23 ++ echo 1.30 1.23 ++ tr ' ' '\n' ++ sort -rV ++ head -n 1 + test 1.30 '!=' 1.30 + rm ./08-assert.yaml + skip_test streaming-with-tls 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=streaming-with-tls + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-with-autoprovisioning-autoscale + '[' streaming-with-autoprovisioning-autoscale '!=' _build ']' + cd .. + rm -rf streaming-with-tls + warning 'streaming-with-tls: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: streaming-with-tls: This test is flaky in Prow CI\e[0m' WAR: streaming-with-tls: This test is flaky in Prow CI + skip_test streaming-simple 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=streaming-simple + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build + '[' _build '!=' _build ']' + rm -rf streaming-simple + warning 'streaming-simple: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: streaming-simple: This test is flaky in Prow CI\e[0m' WAR: streaming-simple: This test is flaky in Prow CI make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running streaming E2E tests' Running streaming E2E tests + cd tests/e2e/streaming/_build + set +e + KUBECONFIG=/tmp/kubeconfig-3197734706 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 2 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/streaming-with-autoprovisioning-autoscale === PAUSE kuttl/harness/streaming-with-autoprovisioning-autoscale === CONT kuttl/harness/artifacts logger.go:42: 07:45:07 | artifacts | Creating namespace: kuttl-test-certain-pegasus logger.go:42: 07:45:07 | artifacts | artifacts events from ns kuttl-test-certain-pegasus: logger.go:42: 07:45:07 | artifacts | Deleting namespace: kuttl-test-certain-pegasus === CONT kuttl/harness/streaming-with-autoprovisioning-autoscale logger.go:42: 07:45:13 | streaming-with-autoprovisioning-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:45:13 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_0.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:45:13 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_1.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:45:13 | streaming-with-autoprovisioning-autoscale | Creating namespace: kuttl-test-fresh-kangaroo logger.go:42: 07:45:13 | streaming-with-autoprovisioning-autoscale/1-install | starting test step 1-install logger.go:42: 07:45:13 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc create sa deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 07:45:13 | streaming-with-autoprovisioning-autoscale/1-install | serviceaccount/deploy-elasticsearch created logger.go:42: 07:45:13 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc adm policy add-scc-to-user privileged -z deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 07:45:13 | streaming-with-autoprovisioning-autoscale/1-install | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:privileged added: "deploy-elasticsearch" logger.go:42: 07:45:13 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 6] logger.go:42: 07:45:19 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_0.yml -n $NAMESPACE] logger.go:42: 07:45:19 | streaming-with-autoprovisioning-autoscale/1-install | statefulset.apps/elasticsearch created logger.go:42: 07:45:20 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 3] logger.go:42: 07:45:23 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_1.yml -n $NAMESPACE] logger.go:42: 07:45:23 | streaming-with-autoprovisioning-autoscale/1-install | service/elasticsearch created logger.go:42: 07:45:40 | streaming-with-autoprovisioning-autoscale/1-install | test step completed 1-install logger.go:42: 07:45:40 | streaming-with-autoprovisioning-autoscale/2-install | starting test step 2-install logger.go:42: 07:45:41 | streaming-with-autoprovisioning-autoscale/2-install | Jaeger:kuttl-test-fresh-kangaroo/auto-provisioned created logger.go:42: 07:45:41 | streaming-with-autoprovisioning-autoscale/2-install | test step completed 2-install logger.go:42: 07:45:41 | streaming-with-autoprovisioning-autoscale/3- | starting test step 3- logger.go:42: 07:46:28 | streaming-with-autoprovisioning-autoscale/3- | test step completed 3- logger.go:42: 07:46:28 | streaming-with-autoprovisioning-autoscale/4- | starting test step 4- logger.go:42: 07:47:12 | streaming-with-autoprovisioning-autoscale/4- | test step completed 4- logger.go:42: 07:47:12 | streaming-with-autoprovisioning-autoscale/5- | starting test step 5- logger.go:42: 07:47:36 | streaming-with-autoprovisioning-autoscale/5- | test step completed 5- logger.go:42: 07:47:36 | streaming-with-autoprovisioning-autoscale/6- | starting test step 6- logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale/6- | test step completed 6- logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale/7- | starting test step 7- logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale/7- | test step completed 7- logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | streaming-with-autoprovisioning-autoscale events from ns kuttl-test-fresh-kangaroo: logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:45:20 +0000 UTC Normal Pod elasticsearch-0 Binding Scheduled Successfully assigned kuttl-test-fresh-kangaroo/elasticsearch-0 to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:45:20 +0000 UTC Normal Pod elasticsearch-0 AddedInterface Add eth0 [10.131.0.59/23] from ovn-kubernetes multus logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:45:20 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulling Pulling image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:45:20 +0000 UTC Normal StatefulSet.apps elasticsearch SuccessfulCreate create Pod elasticsearch-0 in StatefulSet elasticsearch successful statefulset-controller logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:45:27 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulled Successfully pulled image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" in 7.099s (7.099s including waiting). Image size: 758467647 bytes. kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:45:27 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:45:27 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:45:35 +0000 UTC Warning Pod elasticsearch-0.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Get "http://10.131.0.59:9200/": dial tcp 10.131.0.59:9200: connect: connection refused kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:45:49 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-zookeeper NoPods No matching pods found controllermanager logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:45:49 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:45:50 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:45:50 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-fresh-kangaroo/data-auto-provisioned-zookeeper-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-68bc698f7f-lkmgx_6ee59683-c4db-4c00-8e5a-5112a2289b99 logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:45:52 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ProvisioningSucceeded Successfully provisioned volume pvc-4b78e0b6-4ace-417f-aa2f-ef84be0802d8 ebs.csi.aws.com_aws-ebs-csi-driver-controller-68bc698f7f-lkmgx_6ee59683-c4db-4c00-8e5a-5112a2289b99 logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:45:53 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 Binding Scheduled Successfully assigned kuttl-test-fresh-kangaroo/auto-provisioned-zookeeper-0 to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:45:55 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-4b78e0b6-4ace-417f-aa2f-ef84be0802d8" attachdetach-controller logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:45:58 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 AddedInterface Add eth0 [10.129.2.71/23] from ovn-kubernetes multus logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:45:58 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulling Pulling image "registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:c9a60d03827466ee37dd4aff6803eda2c1d65cb2508cb57d13e1d47f982e5a20" kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:46:09 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulled Successfully pulled image "registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:c9a60d03827466ee37dd4aff6803eda2c1d65cb2508cb57d13e1d47f982e5a20" in 11.122s (11.122s including waiting). Image size: 616797249 bytes. kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:46:10 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Created Created container zookeeper kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:46:10 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Started Started container zookeeper kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:46:30 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:46:31 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-kafka NoPods No matching pods found controllermanager logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:46:31 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:46:31 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-fresh-kangaroo/data-0-auto-provisioned-kafka-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-68bc698f7f-lkmgx_6ee59683-c4db-4c00-8e5a-5112a2289b99 logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:46:33 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ProvisioningSucceeded Successfully provisioned volume pvc-5384f692-3bdb-48be-9984-2017d14ff1ca ebs.csi.aws.com_aws-ebs-csi-driver-controller-68bc698f7f-lkmgx_6ee59683-c4db-4c00-8e5a-5112a2289b99 logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:46:34 +0000 UTC Normal Pod auto-provisioned-kafka-0 Binding Scheduled Successfully assigned kuttl-test-fresh-kangaroo/auto-provisioned-kafka-0 to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:46:36 +0000 UTC Normal Pod auto-provisioned-kafka-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-5384f692-3bdb-48be-9984-2017d14ff1ca" attachdetach-controller logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:46:42 +0000 UTC Normal Pod auto-provisioned-kafka-0 AddedInterface Add eth0 [10.131.0.60/23] from ovn-kubernetes multus logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:46:42 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Pulling Pulling image "registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:c9a60d03827466ee37dd4aff6803eda2c1d65cb2508cb57d13e1d47f982e5a20" kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:46:50 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Pulled Successfully pulled image "registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:c9a60d03827466ee37dd4aff6803eda2c1d65cb2508cb57d13e1d47f982e5a20" in 7.923s (7.923s including waiting). Image size: 616797249 bytes. kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:46:50 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Created Created container kafka kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:46:50 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Started Started container kafka kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:15 +0000 UTC Normal Pod auto-provisioned-entity-operator-6f585bf568-wqfbn Binding Scheduled Successfully assigned kuttl-test-fresh-kangaroo/auto-provisioned-entity-operator-6f585bf568-wqfbn to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:15 +0000 UTC Normal Pod auto-provisioned-entity-operator-6f585bf568-wqfbn AddedInterface Add eth0 [10.128.2.82/23] from ovn-kubernetes multus logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:15 +0000 UTC Normal Pod auto-provisioned-entity-operator-6f585bf568-wqfbn.spec.containers{topic-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:07a9540a8b906c6d52e6a9684cfb838aadf9849cc24b1d80218ea3ad5545cb5a" already present on machine kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:15 +0000 UTC Normal Pod auto-provisioned-entity-operator-6f585bf568-wqfbn.spec.containers{topic-operator} Created Created container topic-operator kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:15 +0000 UTC Normal Pod auto-provisioned-entity-operator-6f585bf568-wqfbn.spec.containers{topic-operator} Started Started container topic-operator kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:15 +0000 UTC Normal Pod auto-provisioned-entity-operator-6f585bf568-wqfbn.spec.containers{user-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:07a9540a8b906c6d52e6a9684cfb838aadf9849cc24b1d80218ea3ad5545cb5a" already present on machine kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:15 +0000 UTC Normal Pod auto-provisioned-entity-operator-6f585bf568-wqfbn.spec.containers{user-operator} Created Created container user-operator kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:15 +0000 UTC Normal Pod auto-provisioned-entity-operator-6f585bf568-wqfbn.spec.containers{user-operator} Started Started container user-operator kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:15 +0000 UTC Normal ReplicaSet.apps auto-provisioned-entity-operator-6f585bf568 SuccessfulCreate Created pod: auto-provisioned-entity-operator-6f585bf568-wqfbn replicaset-controller logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:15 +0000 UTC Normal Deployment.apps auto-provisioned-entity-operator ScalingReplicaSet Scaled up replica set auto-provisioned-entity-operator-6f585bf568 to 1 deployment-controller logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:37 +0000 UTC Normal ReplicaSet.apps auto-provisioned-collector-7cd7db7b58 SuccessfulCreate Created pod: auto-provisioned-collector-7cd7db7b58-76bdw replicaset-controller logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:37 +0000 UTC Normal Deployment.apps auto-provisioned-collector ScalingReplicaSet Scaled up replica set auto-provisioned-collector-7cd7db7b58 to 1 deployment-controller logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:37 +0000 UTC Normal Deployment.apps auto-provisioned-query ScalingReplicaSet Scaled up replica set auto-provisioned-query-7778b4cd74 to 1 deployment-controller logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:38 +0000 UTC Normal Pod auto-provisioned-collector-7cd7db7b58-76bdw Binding Scheduled Successfully assigned kuttl-test-fresh-kangaroo/auto-provisioned-collector-7cd7db7b58-76bdw to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:38 +0000 UTC Normal Pod auto-provisioned-collector-7cd7db7b58-76bdw AddedInterface Add eth0 [10.129.2.72/23] from ovn-kubernetes multus logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:38 +0000 UTC Normal Pod auto-provisioned-collector-7cd7db7b58-76bdw.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:38 +0000 UTC Normal Pod auto-provisioned-collector-7cd7db7b58-76bdw.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:38 +0000 UTC Normal Pod auto-provisioned-collector-7cd7db7b58-76bdw.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:38 +0000 UTC Normal Pod auto-provisioned-ingester-5cd9574d99-pz674 Binding Scheduled Successfully assigned kuttl-test-fresh-kangaroo/auto-provisioned-ingester-5cd9574d99-pz674 to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:38 +0000 UTC Normal Pod auto-provisioned-ingester-5cd9574d99-pz674 AddedInterface Add eth0 [10.131.0.61/23] from ovn-kubernetes multus logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:38 +0000 UTC Normal Pod auto-provisioned-ingester-5cd9574d99-pz674.spec.containers{jaeger-ingester} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:e01a99dddffe65385d77dd6692558aa9a47df36f8b2a8d141b6ad561139981b2" kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:38 +0000 UTC Normal ReplicaSet.apps auto-provisioned-ingester-5cd9574d99 SuccessfulCreate Created pod: auto-provisioned-ingester-5cd9574d99-pz674 replicaset-controller logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:38 +0000 UTC Normal Deployment.apps auto-provisioned-ingester ScalingReplicaSet Scaled up replica set auto-provisioned-ingester-5cd9574d99 to 1 deployment-controller logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:38 +0000 UTC Normal Pod auto-provisioned-query-7778b4cd74-f92bm Binding Scheduled Successfully assigned kuttl-test-fresh-kangaroo/auto-provisioned-query-7778b4cd74-f92bm to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:38 +0000 UTC Normal Pod auto-provisioned-query-7778b4cd74-f92bm AddedInterface Add eth0 [10.129.2.73/23] from ovn-kubernetes multus logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:38 +0000 UTC Normal Pod auto-provisioned-query-7778b4cd74-f92bm.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:38 +0000 UTC Normal Pod auto-provisioned-query-7778b4cd74-f92bm.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:38 +0000 UTC Normal Pod auto-provisioned-query-7778b4cd74-f92bm.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:38 +0000 UTC Normal Pod auto-provisioned-query-7778b4cd74-f92bm.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:38 +0000 UTC Normal Pod auto-provisioned-query-7778b4cd74-f92bm.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:38 +0000 UTC Normal Pod auto-provisioned-query-7778b4cd74-f92bm.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:38 +0000 UTC Normal Pod auto-provisioned-query-7778b4cd74-f92bm.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:38 +0000 UTC Normal Pod auto-provisioned-query-7778b4cd74-f92bm.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:38 +0000 UTC Normal Pod auto-provisioned-query-7778b4cd74-f92bm.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:38 +0000 UTC Normal ReplicaSet.apps auto-provisioned-query-7778b4cd74 SuccessfulCreate Created pod: auto-provisioned-query-7778b4cd74-f92bm replicaset-controller logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:46 +0000 UTC Normal Pod auto-provisioned-ingester-5cd9574d99-pz674.spec.containers{jaeger-ingester} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:e01a99dddffe65385d77dd6692558aa9a47df36f8b2a8d141b6ad561139981b2" in 7.435s (7.435s including waiting). Image size: 139765359 bytes. kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:46 +0000 UTC Normal Pod auto-provisioned-ingester-5cd9574d99-pz674.spec.containers{jaeger-ingester} Created Created container jaeger-ingester kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:46 +0000 UTC Normal Pod auto-provisioned-ingester-5cd9574d99-pz674.spec.containers{jaeger-ingester} Started Started container jaeger-ingester kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:47 +0000 UTC Warning Pod auto-provisioned-ingester-5cd9574d99-pz674.spec.containers{jaeger-ingester} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:53 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:53 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:53 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:53 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-ingester FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:53 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-ingester FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | 2025-01-20 07:47:53 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-ingester FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:47:58 | streaming-with-autoprovisioning-autoscale | Deleting namespace: kuttl-test-fresh-kangaroo === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (189.87s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.13s) --- PASS: kuttl/harness/streaming-with-autoprovisioning-autoscale (183.54s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name streaming --report --output /logs/artifacts/streaming.xml ./artifacts/kuttl-report.xml time="2025-01-20T07:48:17Z" level=debug msg="Setting a new name for the test suites" time="2025-01-20T07:48:17Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-01-20T07:48:17Z" level=debug msg="normalizing test case names" time="2025-01-20T07:48:17Z" level=debug msg="streaming/artifacts -> streaming_artifacts" time="2025-01-20T07:48:17Z" level=debug msg="streaming/streaming-with-autoprovisioning-autoscale -> streaming_streaming_with_autoprovisioning_autoscale" +-----------------------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------------------+--------+ | streaming_artifacts | passed | | streaming_streaming_with_autoprovisioning_autoscale | passed | +-----------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh ui false true + '[' 3 -ne 3 ']' + test_suite_name=ui + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/ui.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-ui make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true ./tests/e2e/ui/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-15-030107 True False 54m Cluster version is 4.18.0-0.nightly-2025-01-15-030107' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-01-15-030107 True False 54m Cluster version is 4.18.0-0.nightly-2025-01-15-030107' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/ui/render.sh ++ export SUITE_DIR=./tests/e2e/ui ++ SUITE_DIR=./tests/e2e/ui ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/ui ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test allinone + '[' 1 -ne 1 ']' + test_name=allinone + echo =========================================================================== =========================================================================== + info 'Rendering files for test allinone' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test allinone\e[0m' Rendering files for test allinone + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build + '[' _build '!=' _build ']' + mkdir -p allinone + cd allinone + export GET_URL_COMMAND + export URL + export JAEGER_NAME=all-in-one-ui + JAEGER_NAME=all-in-one-ui + '[' true = true ']' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./01-curl.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./04-test-ui-config.yaml + start_test production + '[' 1 -ne 1 ']' + test_name=production + echo =========================================================================== =========================================================================== + info 'Rendering files for test production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test production\e[0m' Rendering files for test production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build/allinone + '[' allinone '!=' _build ']' + cd .. + mkdir -p production + cd production + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + [[ true = true ]] + [[ true = true ]] + render_install_jaeger production-ui production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + '[' true = true ']' + INSECURE=true + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-forbbiden-access.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-curl.yaml + INSECURE=true + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./05-check-disabled-security.yaml + ASSERT_PRESENT=false + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./06-check-NO-gaID.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./08-check-gaID.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running ui E2E tests' Running ui E2E tests + cd tests/e2e/ui/_build + set +e + KUBECONFIG=/tmp/kubeconfig-3197734706 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 3 tests === RUN kuttl/harness === RUN kuttl/harness/allinone === PAUSE kuttl/harness/allinone === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/production === PAUSE kuttl/harness/production === CONT kuttl/harness/allinone logger.go:42: 07:48:23 | allinone | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:48:23 | allinone | Creating namespace: kuttl-test-communal-warthog logger.go:42: 07:48:23 | allinone/0-install | starting test step 0-install logger.go:42: 07:48:23 | allinone/0-install | Jaeger:kuttl-test-communal-warthog/all-in-one-ui created logger.go:42: 07:48:28 | allinone/0-install | test step completed 0-install logger.go:42: 07:48:28 | allinone/1-curl | starting test step 1-curl logger.go:42: 07:48:28 | allinone/1-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 07:48:28 | allinone/1-curl | Checking the Ingress host value was populated logger.go:42: 07:48:28 | allinone/1-curl | Try number 0 logger.go:42: 07:48:28 | allinone/1-curl | Hostname is all-in-one-ui-kuttl-test-communal-warthog.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com logger.go:42: 07:48:28 | allinone/1-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE all-in-one-ui] logger.go:42: 07:48:28 | allinone/1-curl | Checking an expected HTTP response logger.go:42: 07:48:28 | allinone/1-curl | Running in OpenShift logger.go:42: 07:48:28 | allinone/1-curl | User not provided. Getting the token... logger.go:42: 07:48:29 | allinone/1-curl | Warning: resource jaegers/all-in-one-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:48:35 | allinone/1-curl | Try number 1/30 the https://all-in-one-ui-kuttl-test-communal-warthog.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:48:35 | allinone/1-curl | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:48:35 | allinone/1-curl | Try number 2/30 the https://all-in-one-ui-kuttl-test-communal-warthog.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:48:56 | allinone/1-curl | HTTP response is 503. 200 expected. Waiting 10 s logger.go:42: 07:49:06 | allinone/1-curl | Try number 3/30 the https://all-in-one-ui-kuttl-test-communal-warthog.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:49:06 | allinone/1-curl | curl response asserted properly logger.go:42: 07:49:06 | allinone/1-curl | test step completed 1-curl logger.go:42: 07:49:06 | allinone/2-delete | starting test step 2-delete logger.go:42: 07:49:06 | allinone/2-delete | Jaeger:kuttl-test-communal-warthog/all-in-one-ui created logger.go:42: 07:49:06 | allinone/2-delete | test step completed 2-delete logger.go:42: 07:49:06 | allinone/3-install | starting test step 3-install logger.go:42: 07:49:06 | allinone/3-install | Jaeger:kuttl-test-communal-warthog/all-in-one-ui updated logger.go:42: 07:49:06 | allinone/3-install | test step completed 3-install logger.go:42: 07:49:06 | allinone/4-test-ui-config | starting test step 4-test-ui-config logger.go:42: 07:49:06 | allinone/4-test-ui-config | running command: [./ensure-ingress-host.sh] logger.go:42: 07:49:06 | allinone/4-test-ui-config | Checking the Ingress host value was populated logger.go:42: 07:49:06 | allinone/4-test-ui-config | Try number 0 logger.go:42: 07:49:06 | allinone/4-test-ui-config | error: error executing jsonpath "{.items[0].status.ingress[0].host}": Error executing template: array index out of bounds: index 0, length 0. Printing more information for debugging the template: logger.go:42: 07:49:06 | allinone/4-test-ui-config | template was: logger.go:42: 07:49:06 | allinone/4-test-ui-config | {.items[0].status.ingress[0].host} logger.go:42: 07:49:06 | allinone/4-test-ui-config | object given to jsonpath engine was: logger.go:42: 07:49:06 | allinone/4-test-ui-config | map[string]interface {}{"apiVersion":"v1", "items":[]interface {}{}, "kind":"List", "metadata":map[string]interface {}{"resourceVersion":""}} logger.go:42: 07:49:06 | allinone/4-test-ui-config | logger.go:42: 07:49:06 | allinone/4-test-ui-config | logger.go:42: 07:49:16 | allinone/4-test-ui-config | Try number 1 logger.go:42: 07:49:16 | allinone/4-test-ui-config | Hostname is all-in-one-ui-kuttl-test-communal-warthog.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com logger.go:42: 07:49:16 | allinone/4-test-ui-config | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 07:49:17 | allinone/4-test-ui-config | time="2025-01-20T07:49:17Z" level=info msg="Querying https://all-in-one-ui-kuttl-test-communal-warthog.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com/search..." logger.go:42: 07:49:17 | allinone/4-test-ui-config | time="2025-01-20T07:49:17Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 07:49:17 | allinone/4-test-ui-config | time="2025-01-20T07:49:17Z" level=info msg="Polling to https://all-in-one-ui-kuttl-test-communal-warthog.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com/search" logger.go:42: 07:49:17 | allinone/4-test-ui-config | time="2025-01-20T07:49:17Z" level=info msg="Doing request number 0" logger.go:42: 07:49:17 | allinone/4-test-ui-config | time="2025-01-20T07:49:17Z" level=info msg="Content found and asserted!" logger.go:42: 07:49:17 | allinone/4-test-ui-config | time="2025-01-20T07:49:17Z" level=info msg="Success!" logger.go:42: 07:49:17 | allinone/4-test-ui-config | test step completed 4-test-ui-config logger.go:42: 07:49:17 | allinone | allinone events from ns kuttl-test-communal-warthog: logger.go:42: 07:49:17 | allinone | 2025-01-20 07:48:27 +0000 UTC Normal Pod all-in-one-ui-748d95874b-zrvff Binding Scheduled Successfully assigned kuttl-test-communal-warthog/all-in-one-ui-748d95874b-zrvff to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:49:17 | allinone | 2025-01-20 07:48:27 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-748d95874b SuccessfulCreate Created pod: all-in-one-ui-748d95874b-zrvff replicaset-controller logger.go:42: 07:49:17 | allinone | 2025-01-20 07:48:27 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-748d95874b to 1 deployment-controller logger.go:42: 07:49:17 | allinone | 2025-01-20 07:48:28 +0000 UTC Warning Pod all-in-one-ui-748d95874b-zrvff FailedMount MountVolume.SetUp failed for volume "all-in-one-ui-collector-tls-config-volume" : secret "all-in-one-ui-collector-headless-tls" not found kubelet logger.go:42: 07:49:17 | allinone | 2025-01-20 07:48:29 +0000 UTC Normal Pod all-in-one-ui-748d95874b-zrvff AddedInterface Add eth0 [10.131.0.62/23] from ovn-kubernetes multus logger.go:42: 07:49:17 | allinone | 2025-01-20 07:48:29 +0000 UTC Normal Pod all-in-one-ui-748d95874b-zrvff.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:49:17 | allinone | 2025-01-20 07:48:29 +0000 UTC Normal Pod all-in-one-ui-748d95874b-zrvff.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:49:17 | allinone | 2025-01-20 07:48:29 +0000 UTC Normal Pod all-in-one-ui-748d95874b-zrvff.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:49:17 | allinone | 2025-01-20 07:48:29 +0000 UTC Normal Pod all-in-one-ui-748d95874b-zrvff.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:49:17 | allinone | 2025-01-20 07:48:29 +0000 UTC Normal Pod all-in-one-ui-748d95874b-zrvff.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:49:17 | allinone | 2025-01-20 07:48:29 +0000 UTC Normal Pod all-in-one-ui-748d95874b-zrvff.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:49:17 | allinone | 2025-01-20 07:48:33 +0000 UTC Normal Pod all-in-one-ui-748d95874b-zrvff.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:49:17 | allinone | 2025-01-20 07:48:33 +0000 UTC Normal Pod all-in-one-ui-748d95874b-zrvff.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:49:17 | allinone | 2025-01-20 07:48:33 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-748d95874b SuccessfulDelete Deleted pod: all-in-one-ui-748d95874b-zrvff replicaset-controller logger.go:42: 07:49:17 | allinone | 2025-01-20 07:48:33 +0000 UTC Normal Pod all-in-one-ui-f7c58fc58-6267h Binding Scheduled Successfully assigned kuttl-test-communal-warthog/all-in-one-ui-f7c58fc58-6267h to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:49:17 | allinone | 2025-01-20 07:48:33 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-f7c58fc58 SuccessfulCreate Created pod: all-in-one-ui-f7c58fc58-6267h replicaset-controller logger.go:42: 07:49:17 | allinone | 2025-01-20 07:48:33 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled down replica set all-in-one-ui-748d95874b to 0 from 1 deployment-controller logger.go:42: 07:49:17 | allinone | 2025-01-20 07:48:33 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-f7c58fc58 to 1 deployment-controller logger.go:42: 07:49:17 | allinone | 2025-01-20 07:48:34 +0000 UTC Normal Pod all-in-one-ui-f7c58fc58-6267h AddedInterface Add eth0 [10.131.0.63/23] from ovn-kubernetes multus logger.go:42: 07:49:17 | allinone | 2025-01-20 07:48:34 +0000 UTC Normal Pod all-in-one-ui-f7c58fc58-6267h.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:49:17 | allinone | 2025-01-20 07:48:34 +0000 UTC Normal Pod all-in-one-ui-f7c58fc58-6267h.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:49:17 | allinone | 2025-01-20 07:48:34 +0000 UTC Normal Pod all-in-one-ui-f7c58fc58-6267h.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:49:17 | allinone | 2025-01-20 07:48:34 +0000 UTC Normal Pod all-in-one-ui-f7c58fc58-6267h.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:49:17 | allinone | 2025-01-20 07:48:34 +0000 UTC Normal Pod all-in-one-ui-f7c58fc58-6267h.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:49:17 | allinone | 2025-01-20 07:48:34 +0000 UTC Normal Pod all-in-one-ui-f7c58fc58-6267h.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:49:17 | allinone | 2025-01-20 07:49:06 +0000 UTC Normal Pod all-in-one-ui-f7c58fc58-6267h.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:49:17 | allinone | 2025-01-20 07:49:06 +0000 UTC Normal Pod all-in-one-ui-f7c58fc58-6267h.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:49:17 | allinone | 2025-01-20 07:49:12 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-78bf594b6b to 1 deployment-controller logger.go:42: 07:49:17 | allinone | 2025-01-20 07:49:13 +0000 UTC Normal Pod all-in-one-ui-78bf594b6b-c22kx Binding Scheduled Successfully assigned kuttl-test-communal-warthog/all-in-one-ui-78bf594b6b-c22kx to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:49:17 | allinone | 2025-01-20 07:49:13 +0000 UTC Normal Pod all-in-one-ui-78bf594b6b-c22kx AddedInterface Add eth0 [10.131.0.64/23] from ovn-kubernetes multus logger.go:42: 07:49:17 | allinone | 2025-01-20 07:49:13 +0000 UTC Normal Pod all-in-one-ui-78bf594b6b-c22kx.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:49:17 | allinone | 2025-01-20 07:49:13 +0000 UTC Normal Pod all-in-one-ui-78bf594b6b-c22kx.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:49:17 | allinone | 2025-01-20 07:49:13 +0000 UTC Normal Pod all-in-one-ui-78bf594b6b-c22kx.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:49:17 | allinone | 2025-01-20 07:49:13 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-78bf594b6b SuccessfulCreate Created pod: all-in-one-ui-78bf594b6b-c22kx replicaset-controller logger.go:42: 07:49:17 | allinone | Deleting namespace: kuttl-test-communal-warthog === CONT kuttl/harness/production logger.go:42: 07:49:24 | production | Ignoring add-tracking-id.yaml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:49:24 | production | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:49:24 | production | Creating namespace: kuttl-test-viable-parrot logger.go:42: 07:49:24 | production/1-install | starting test step 1-install logger.go:42: 07:49:24 | production/1-install | Jaeger:kuttl-test-viable-parrot/production-ui created logger.go:42: 07:50:03 | production/1-install | test step completed 1-install logger.go:42: 07:50:03 | production/2-check-forbbiden-access | starting test step 2-check-forbbiden-access logger.go:42: 07:50:03 | production/2-check-forbbiden-access | running command: [./ensure-ingress-host.sh] logger.go:42: 07:50:03 | production/2-check-forbbiden-access | Checking the Ingress host value was populated logger.go:42: 07:50:03 | production/2-check-forbbiden-access | Try number 0 logger.go:42: 07:50:03 | production/2-check-forbbiden-access | Hostname is production-ui-kuttl-test-viable-parrot.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com logger.go:42: 07:50:03 | production/2-check-forbbiden-access | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE production-ui] logger.go:42: 07:50:03 | production/2-check-forbbiden-access | Checking an expected HTTP response logger.go:42: 07:50:03 | production/2-check-forbbiden-access | Running in OpenShift logger.go:42: 07:50:03 | production/2-check-forbbiden-access | Not using any secret logger.go:42: 07:50:03 | production/2-check-forbbiden-access | Try number 1/30 the https://production-ui-kuttl-test-viable-parrot.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:50:03 | production/2-check-forbbiden-access | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:50:03 | production/2-check-forbbiden-access | Try number 2/30 the https://production-ui-kuttl-test-viable-parrot.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:50:03 | production/2-check-forbbiden-access | HTTP response is 503. 403 expected. Waiting 10 s logger.go:42: 07:50:13 | production/2-check-forbbiden-access | Try number 3/30 the https://production-ui-kuttl-test-viable-parrot.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:50:13 | production/2-check-forbbiden-access | curl response asserted properly logger.go:42: 07:50:13 | production/2-check-forbbiden-access | test step completed 2-check-forbbiden-access logger.go:42: 07:50:13 | production/3-curl | starting test step 3-curl logger.go:42: 07:50:13 | production/3-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 07:50:13 | production/3-curl | Checking the Ingress host value was populated logger.go:42: 07:50:13 | production/3-curl | Try number 0 logger.go:42: 07:50:14 | production/3-curl | Hostname is production-ui-kuttl-test-viable-parrot.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com logger.go:42: 07:50:14 | production/3-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 07:50:14 | production/3-curl | Checking an expected HTTP response logger.go:42: 07:50:14 | production/3-curl | Running in OpenShift logger.go:42: 07:50:14 | production/3-curl | User not provided. Getting the token... logger.go:42: 07:50:15 | production/3-curl | Warning: resource jaegers/production-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:50:21 | production/3-curl | Try number 1/30 the https://production-ui-kuttl-test-viable-parrot.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:50:21 | production/3-curl | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:50:21 | production/3-curl | Try number 2/30 the https://production-ui-kuttl-test-viable-parrot.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:50:21 | production/3-curl | HTTP response is 503. 200 expected. Waiting 10 s logger.go:42: 07:50:31 | production/3-curl | Try number 3/30 the https://production-ui-kuttl-test-viable-parrot.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:50:31 | production/3-curl | curl response asserted properly logger.go:42: 07:50:31 | production/3-curl | test step completed 3-curl logger.go:42: 07:50:31 | production/4-install | starting test step 4-install logger.go:42: 07:50:32 | production/4-install | Jaeger:kuttl-test-viable-parrot/production-ui updated logger.go:42: 07:50:32 | production/4-install | test step completed 4-install logger.go:42: 07:50:32 | production/5-check-disabled-security | starting test step 5-check-disabled-security logger.go:42: 07:50:32 | production/5-check-disabled-security | running command: [./ensure-ingress-host.sh] logger.go:42: 07:50:32 | production/5-check-disabled-security | Checking the Ingress host value was populated logger.go:42: 07:50:32 | production/5-check-disabled-security | Try number 0 logger.go:42: 07:50:32 | production/5-check-disabled-security | Hostname is production-ui-kuttl-test-viable-parrot.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com logger.go:42: 07:50:32 | production/5-check-disabled-security | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 07:50:32 | production/5-check-disabled-security | Checking an expected HTTP response logger.go:42: 07:50:32 | production/5-check-disabled-security | Running in OpenShift logger.go:42: 07:50:32 | production/5-check-disabled-security | Not using any secret logger.go:42: 07:50:32 | production/5-check-disabled-security | Try number 1/30 the https://production-ui-kuttl-test-viable-parrot.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:50:32 | production/5-check-disabled-security | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:50:32 | production/5-check-disabled-security | Try number 2/30 the https://production-ui-kuttl-test-viable-parrot.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:50:32 | production/5-check-disabled-security | HTTP response is 403. 200 expected. Waiting 10 s logger.go:42: 07:50:42 | production/5-check-disabled-security | Try number 3/30 the https://production-ui-kuttl-test-viable-parrot.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:50:42 | production/5-check-disabled-security | curl response asserted properly logger.go:42: 07:50:42 | production/5-check-disabled-security | test step completed 5-check-disabled-security logger.go:42: 07:50:42 | production/6-check-NO-gaID | starting test step 6-check-NO-gaID logger.go:42: 07:50:42 | production/6-check-NO-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 07:50:42 | production/6-check-NO-gaID | Checking the Ingress host value was populated logger.go:42: 07:50:42 | production/6-check-NO-gaID | Try number 0 logger.go:42: 07:50:42 | production/6-check-NO-gaID | Hostname is production-ui-kuttl-test-viable-parrot.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com logger.go:42: 07:50:42 | production/6-check-NO-gaID | running command: [sh -c ASSERT_PRESENT=false EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 07:50:43 | production/6-check-NO-gaID | time="2025-01-20T07:50:43Z" level=info msg="Querying https://production-ui-kuttl-test-viable-parrot.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com/search..." logger.go:42: 07:50:43 | production/6-check-NO-gaID | time="2025-01-20T07:50:43Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 07:50:43 | production/6-check-NO-gaID | time="2025-01-20T07:50:43Z" level=info msg="Polling to https://production-ui-kuttl-test-viable-parrot.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com/search" logger.go:42: 07:50:43 | production/6-check-NO-gaID | time="2025-01-20T07:50:43Z" level=info msg="Doing request number 0" logger.go:42: 07:50:43 | production/6-check-NO-gaID | time="2025-01-20T07:50:43Z" level=info msg="Content not found and asserted it was not found!" logger.go:42: 07:50:43 | production/6-check-NO-gaID | time="2025-01-20T07:50:43Z" level=info msg="Success!" logger.go:42: 07:50:43 | production/6-check-NO-gaID | test step completed 6-check-NO-gaID logger.go:42: 07:50:43 | production/7-add-tracking-id | starting test step 7-add-tracking-id logger.go:42: 07:50:43 | production/7-add-tracking-id | running command: [sh -c kubectl apply -f add-tracking-id.yaml -n $NAMESPACE] logger.go:42: 07:50:43 | production/7-add-tracking-id | jaeger.jaegertracing.io/production-ui configured logger.go:42: 07:50:43 | production/7-add-tracking-id | test step completed 7-add-tracking-id logger.go:42: 07:50:43 | production/8-check-gaID | starting test step 8-check-gaID logger.go:42: 07:50:43 | production/8-check-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 07:50:43 | production/8-check-gaID | Checking the Ingress host value was populated logger.go:42: 07:50:43 | production/8-check-gaID | Try number 0 logger.go:42: 07:50:43 | production/8-check-gaID | Hostname is production-ui-kuttl-test-viable-parrot.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com logger.go:42: 07:50:43 | production/8-check-gaID | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 07:50:44 | production/8-check-gaID | time="2025-01-20T07:50:44Z" level=info msg="Querying https://production-ui-kuttl-test-viable-parrot.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com/search..." logger.go:42: 07:50:44 | production/8-check-gaID | time="2025-01-20T07:50:44Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 07:50:44 | production/8-check-gaID | time="2025-01-20T07:50:44Z" level=info msg="Polling to https://production-ui-kuttl-test-viable-parrot.apps.ci-op-9yglk5c7-9210b.cspilp.interop.ccitredhat.com/search" logger.go:42: 07:50:44 | production/8-check-gaID | time="2025-01-20T07:50:44Z" level=info msg="Doing request number 0" logger.go:42: 07:50:44 | production/8-check-gaID | time="2025-01-20T07:50:44Z" level=warning msg="Found: false . Assert: true" logger.go:42: 07:50:44 | production/8-check-gaID | time="2025-01-20T07:50:44Z" level=warning msg="The condition of the test function was not accomplished" logger.go:42: 07:50:52 | production/8-check-gaID | time="2025-01-20T07:50:52Z" level=info msg="Doing request number 1" logger.go:42: 07:51:12 | production/8-check-gaID | time="2025-01-20T07:51:12Z" level=info msg="Content found and asserted!" logger.go:42: 07:51:12 | production/8-check-gaID | time="2025-01-20T07:51:12Z" level=info msg="Success!" logger.go:42: 07:51:12 | production/8-check-gaID | test step completed 8-check-gaID logger.go:42: 07:51:12 | production | production events from ns kuttl-test-viable-parrot: logger.go:42: 07:51:12 | production | 2025-01-20 07:49:32 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestviableparrotproductionui-1-fcd4f6f69 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestviableparrotproductionui-1-fcd4wsswb replicaset-controller logger.go:42: 07:51:12 | production | 2025-01-20 07:49:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestviableparrotproductionui-1-fcd4wsswb Binding Scheduled Successfully assigned kuttl-test-viable-parrot/elasticsearch-cdm-kuttltestviableparrotproductionui-1-fcd4wsswb to ip-10-0-120-49.ec2.internal default-scheduler logger.go:42: 07:51:12 | production | 2025-01-20 07:49:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestviableparrotproductionui-1-fcd4wsswb AddedInterface Add eth0 [10.131.0.65/23] from ovn-kubernetes multus logger.go:42: 07:51:12 | production | 2025-01-20 07:49:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestviableparrotproductionui-1-fcd4wsswb.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:49:32 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestviableparrotproductionui-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestviableparrotproductionui-1-fcd4f6f69 to 1 deployment-controller logger.go:42: 07:51:12 | production | 2025-01-20 07:49:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestviableparrotproductionui-1-fcd4wsswb.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:49:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestviableparrotproductionui-1-fcd4wsswb.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:49:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestviableparrotproductionui-1-fcd4wsswb.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:49:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestviableparrotproductionui-1-fcd4wsswb.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:49:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestviableparrotproductionui-1-fcd4wsswb.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:49:47 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestviableparrotproductionui-1-fcd4wsswb.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:49:59 +0000 UTC Normal Pod production-ui-collector-5b878c474-hfxvb Binding Scheduled Successfully assigned kuttl-test-viable-parrot/production-ui-collector-5b878c474-hfxvb to ip-10-0-37-75.ec2.internal default-scheduler logger.go:42: 07:51:12 | production | 2025-01-20 07:49:59 +0000 UTC Warning Pod production-ui-collector-5b878c474-hfxvb FailedMount MountVolume.SetUp failed for volume "production-ui-collector-tls-config-volume" : secret "production-ui-collector-headless-tls" not found kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:49:59 +0000 UTC Normal ReplicaSet.apps production-ui-collector-5b878c474 SuccessfulCreate Created pod: production-ui-collector-5b878c474-hfxvb replicaset-controller logger.go:42: 07:51:12 | production | 2025-01-20 07:49:59 +0000 UTC Normal Deployment.apps production-ui-collector ScalingReplicaSet Scaled up replica set production-ui-collector-5b878c474 to 1 deployment-controller logger.go:42: 07:51:12 | production | 2025-01-20 07:49:59 +0000 UTC Normal Pod production-ui-query-7558696844-gzkm2 Binding Scheduled Successfully assigned kuttl-test-viable-parrot/production-ui-query-7558696844-gzkm2 to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:51:12 | production | 2025-01-20 07:49:59 +0000 UTC Warning Pod production-ui-query-7558696844-gzkm2 FailedMount MountVolume.SetUp failed for volume "production-ui-ui-oauth-proxy-tls" : secret "production-ui-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:49:59 +0000 UTC Normal ReplicaSet.apps production-ui-query-7558696844 SuccessfulCreate Created pod: production-ui-query-7558696844-gzkm2 replicaset-controller logger.go:42: 07:51:12 | production | 2025-01-20 07:49:59 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-7558696844 to 1 deployment-controller logger.go:42: 07:51:12 | production | 2025-01-20 07:50:00 +0000 UTC Normal Pod production-ui-collector-5b878c474-hfxvb AddedInterface Add eth0 [10.129.2.74/23] from ovn-kubernetes multus logger.go:42: 07:51:12 | production | 2025-01-20 07:50:00 +0000 UTC Normal Pod production-ui-collector-5b878c474-hfxvb.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:00 +0000 UTC Normal Pod production-ui-collector-5b878c474-hfxvb.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:00 +0000 UTC Normal Pod production-ui-collector-5b878c474-hfxvb.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:00 +0000 UTC Normal Pod production-ui-query-7558696844-gzkm2 AddedInterface Add eth0 [10.128.2.83/23] from ovn-kubernetes multus logger.go:42: 07:51:12 | production | 2025-01-20 07:50:00 +0000 UTC Normal Pod production-ui-query-7558696844-gzkm2.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:00 +0000 UTC Normal Pod production-ui-query-7558696844-gzkm2.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:00 +0000 UTC Normal Pod production-ui-query-7558696844-gzkm2.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:00 +0000 UTC Normal Pod production-ui-query-7558696844-gzkm2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:00 +0000 UTC Normal Pod production-ui-query-7558696844-gzkm2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:00 +0000 UTC Normal Pod production-ui-query-7558696844-gzkm2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:00 +0000 UTC Normal Pod production-ui-query-7558696844-gzkm2.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:00 +0000 UTC Normal Pod production-ui-query-7558696844-gzkm2.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:00 +0000 UTC Normal Pod production-ui-query-7558696844-gzkm2.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:17 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:51:12 | production | 2025-01-20 07:50:17 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod production-ui-collector-5b878c474-hfxvb horizontal-pod-autoscaler logger.go:42: 07:51:12 | production | 2025-01-20 07:50:17 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:51:12 | production | 2025-01-20 07:50:17 +0000 UTC Normal Pod production-ui-query-7558696844-gzkm2.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:17 +0000 UTC Normal Pod production-ui-query-7558696844-gzkm2.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:17 +0000 UTC Normal Pod production-ui-query-7558696844-gzkm2.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:17 +0000 UTC Normal ReplicaSet.apps production-ui-query-7558696844 SuccessfulDelete Deleted pod: production-ui-query-7558696844-gzkm2 replicaset-controller logger.go:42: 07:51:12 | production | 2025-01-20 07:50:17 +0000 UTC Normal Pod production-ui-query-f747479d8-srx28 Binding Scheduled Successfully assigned kuttl-test-viable-parrot/production-ui-query-f747479d8-srx28 to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:51:12 | production | 2025-01-20 07:50:17 +0000 UTC Normal ReplicaSet.apps production-ui-query-f747479d8 SuccessfulCreate Created pod: production-ui-query-f747479d8-srx28 replicaset-controller logger.go:42: 07:51:12 | production | 2025-01-20 07:50:17 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-7558696844 to 0 from 1 deployment-controller logger.go:42: 07:51:12 | production | 2025-01-20 07:50:17 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-f747479d8 to 1 deployment-controller logger.go:42: 07:51:12 | production | 2025-01-20 07:50:18 +0000 UTC Normal Pod production-ui-query-f747479d8-srx28 AddedInterface Add eth0 [10.128.2.84/23] from ovn-kubernetes multus logger.go:42: 07:51:12 | production | 2025-01-20 07:50:18 +0000 UTC Normal Pod production-ui-query-f747479d8-srx28.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:18 +0000 UTC Normal Pod production-ui-query-f747479d8-srx28.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:18 +0000 UTC Normal Pod production-ui-query-f747479d8-srx28.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:18 +0000 UTC Normal Pod production-ui-query-f747479d8-srx28.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:18 +0000 UTC Normal Pod production-ui-query-f747479d8-srx28.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:18 +0000 UTC Normal Pod production-ui-query-f747479d8-srx28.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:18 +0000 UTC Normal Pod production-ui-query-f747479d8-srx28.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:18 +0000 UTC Normal Pod production-ui-query-f747479d8-srx28.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:18 +0000 UTC Normal Pod production-ui-query-f747479d8-srx28.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:32 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:51:12 | production | 2025-01-20 07:50:32 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:51:12 | production | 2025-01-20 07:50:33 +0000 UTC Normal Pod production-ui-query-f747479d8-srx28.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:33 +0000 UTC Normal Pod production-ui-query-f747479d8-srx28.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:33 +0000 UTC Normal Pod production-ui-query-f747479d8-srx28.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:33 +0000 UTC Normal ReplicaSet.apps production-ui-query-f747479d8 SuccessfulDelete Deleted pod: production-ui-query-f747479d8-srx28 replicaset-controller logger.go:42: 07:51:12 | production | 2025-01-20 07:50:33 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-f747479d8 to 0 from 1 deployment-controller logger.go:42: 07:51:12 | production | 2025-01-20 07:50:34 +0000 UTC Normal Pod production-ui-query-568d5b855c-txhrg Binding Scheduled Successfully assigned kuttl-test-viable-parrot/production-ui-query-568d5b855c-txhrg to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:51:12 | production | 2025-01-20 07:50:34 +0000 UTC Normal ReplicaSet.apps production-ui-query-568d5b855c SuccessfulCreate Created pod: production-ui-query-568d5b855c-txhrg replicaset-controller logger.go:42: 07:51:12 | production | 2025-01-20 07:50:34 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-568d5b855c to 1 deployment-controller logger.go:42: 07:51:12 | production | 2025-01-20 07:50:35 +0000 UTC Normal Pod production-ui-query-568d5b855c-txhrg AddedInterface Add eth0 [10.128.2.85/23] from ovn-kubernetes multus logger.go:42: 07:51:12 | production | 2025-01-20 07:50:35 +0000 UTC Normal Pod production-ui-query-568d5b855c-txhrg.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:35 +0000 UTC Normal Pod production-ui-query-568d5b855c-txhrg.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:35 +0000 UTC Normal Pod production-ui-query-568d5b855c-txhrg.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:35 +0000 UTC Normal Pod production-ui-query-568d5b855c-txhrg.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:35 +0000 UTC Normal Pod production-ui-query-568d5b855c-txhrg.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:35 +0000 UTC Normal Pod production-ui-query-568d5b855c-txhrg.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:44 +0000 UTC Normal Pod production-ui-query-568d5b855c-txhrg.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:44 +0000 UTC Normal Pod production-ui-query-568d5b855c-txhrg.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:44 +0000 UTC Normal ReplicaSet.apps production-ui-query-568d5b855c SuccessfulDelete Deleted pod: production-ui-query-568d5b855c-txhrg replicaset-controller logger.go:42: 07:51:12 | production | 2025-01-20 07:50:44 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-568d5b855c to 0 from 1 deployment-controller logger.go:42: 07:51:12 | production | 2025-01-20 07:50:45 +0000 UTC Warning Pod production-ui-query-568d5b855c-txhrg.spec.containers{jaeger-agent} Unhealthy Readiness probe failed: Get "http://10.128.2.85:14271/": dial tcp 10.128.2.85:14271: connect: connection refused kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:45 +0000 UTC Warning Pod production-ui-query-568d5b855c-txhrg.spec.containers{jaeger-query} Unhealthy Readiness probe failed: Get "http://10.128.2.85:16687/": dial tcp 10.128.2.85:16687: connect: connection refused kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:45 +0000 UTC Normal Pod production-ui-query-8679557448-z9phj Binding Scheduled Successfully assigned kuttl-test-viable-parrot/production-ui-query-8679557448-z9phj to ip-10-0-76-173.ec2.internal default-scheduler logger.go:42: 07:51:12 | production | 2025-01-20 07:50:45 +0000 UTC Normal ReplicaSet.apps production-ui-query-8679557448 SuccessfulCreate Created pod: production-ui-query-8679557448-z9phj replicaset-controller logger.go:42: 07:51:12 | production | 2025-01-20 07:50:45 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-8679557448 to 1 deployment-controller logger.go:42: 07:51:12 | production | 2025-01-20 07:50:46 +0000 UTC Normal Pod production-ui-query-8679557448-z9phj AddedInterface Add eth0 [10.128.2.86/23] from ovn-kubernetes multus logger.go:42: 07:51:12 | production | 2025-01-20 07:50:46 +0000 UTC Normal Pod production-ui-query-8679557448-z9phj.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:46 +0000 UTC Normal Pod production-ui-query-8679557448-z9phj.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:46 +0000 UTC Normal Pod production-ui-query-8679557448-z9phj.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:46 +0000 UTC Normal Pod production-ui-query-8679557448-z9phj.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:46 +0000 UTC Normal Pod production-ui-query-8679557448-z9phj.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:46 +0000 UTC Normal Pod production-ui-query-8679557448-z9phj.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:51:12 | production | 2025-01-20 07:50:47 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod production-ui-collector-5b878c474-hfxvb horizontal-pod-autoscaler logger.go:42: 07:51:12 | production | 2025-01-20 07:50:47 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod production-ui-collector-5b878c474-hfxvb horizontal-pod-autoscaler logger.go:42: 07:51:12 | production | Deleting namespace: kuttl-test-viable-parrot === CONT kuttl/harness/artifacts logger.go:42: 07:51:18 | artifacts | Creating namespace: kuttl-test-glad-mink logger.go:42: 07:51:18 | artifacts | artifacts events from ns kuttl-test-glad-mink: logger.go:42: 07:51:18 | artifacts | Deleting namespace: kuttl-test-glad-mink === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (181.36s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/allinone (60.23s) --- PASS: kuttl/harness/production (114.75s) --- PASS: kuttl/harness/artifacts (6.22s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name ui --report --output /logs/artifacts/ui.xml ./artifacts/kuttl-report.xml time="2025-01-20T07:51:25Z" level=debug msg="Setting a new name for the test suites" time="2025-01-20T07:51:25Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-01-20T07:51:25Z" level=debug msg="normalizing test case names" time="2025-01-20T07:51:25Z" level=debug msg="ui/allinone -> ui_allinone" time="2025-01-20T07:51:25Z" level=debug msg="ui/production -> ui_production" time="2025-01-20T07:51:25Z" level=debug msg="ui/artifacts -> ui_artifacts" +---------------+--------+ | NAME | RESULT | +---------------+--------+ | ui_allinone | passed | | ui_production | passed | | ui_artifacts | passed | +---------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests'