% Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 3831 100 3831 0 0 20957 0 --:--:-- --:--:-- --:--:-- 21049 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 9315 100 9315 0 0 37706 0 --:--:-- --:--:-- --:--:-- 37712 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 2607 100 2607 0 0 17233 0 --:--:-- --:--:-- --:--:-- 17264 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 1953 100 1953 0 0 14922 0 --:--:-- --:--:-- --:--:-- 15023 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 350 100 350 0 0 1865 0 --:--:-- --:--:-- --:--:-- 1871 Installing kuttl Try 0... curl -sLo /tmp/jaeger-tests/hack/install/../../bin/kubectl-kuttl https://github.com/kudobuilder/kuttl/releases/download/v0.15.0/kubectl-kuttl_0.15.0_linux_x86_64 KUBECONFIG file is: /tmp/kubeconfig-488946273 for suite in miscellaneous elasticsearch examples generate upgrade sidecar streaming ui; do \ make run-e2e-tests-$suite ; \ done make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh miscellaneous false true + '[' 3 -ne 3 ']' + test_suite_name=miscellaneous + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/miscellaneous.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-miscellaneous make[2]: Entering directory '/tmp/jaeger-tests' SKIP_ES_EXTERNAL=true ./tests/e2e/miscellaneous/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.32 True False 10m Cluster version is 4.15.32' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.32 True False 10m Cluster version is 4.15.32' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/miscellaneous/render.sh ++ export SUITE_DIR=./tests/e2e/miscellaneous ++ SUITE_DIR=./tests/e2e/miscellaneous ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/miscellaneous ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test collector-autoscale + '[' 1 -ne 1 ']' + test_name=collector-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-autoscale\e[0m' Rendering files for test collector-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p collector-autoscale + cd collector-autoscale + jaeger_name=simple-prod + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + ELASTICSEARCH_NODECOUNT=1 + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.resources.requests.memory="200m"' 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.autoscale=true 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.minReplicas=1 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.maxReplicas=2 01-install.yaml + version_lt 1.30 1.23 ++ echo 1.30 1.23 ++ tr ' ' '\n' ++ sort -rV ++ head -n 1 + test 1.30 '!=' 1.30 + rm ./03-assert.yaml + generate_otlp_e2e_tests http + test_protocol=http + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-http\e[0m' Rendering files for test collector-otlp-allinone-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-autoscale + '[' collector-autoscale '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-http + cd collector-otlp-allinone-http + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger http true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-http\e[0m' Rendering files for test collector-otlp-production-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-http + '[' collector-otlp-allinone-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-http + cd collector-otlp-production-http + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger http true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + generate_otlp_e2e_tests grpc + test_protocol=grpc + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-grpc\e[0m' Rendering files for test collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-http + '[' collector-otlp-production-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-grpc + cd collector-otlp-allinone-grpc + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-grpc\e[0m' Rendering files for test collector-otlp-production-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-grpc + '[' collector-otlp-allinone-grpc '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-grpc + cd collector-otlp-production-grpc + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + '[' true = true ']' + skip_test istio 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=istio + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-grpc + '[' collector-otlp-production-grpc '!=' _build ']' + cd .. + rm -rf istio + warning 'istio: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: istio: Test not supported in OpenShift\e[0m' WAR: istio: Test not supported in OpenShift + '[' true = true ']' + skip_test outside-cluster 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=outside-cluster + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + rm -rf outside-cluster + warning 'outside-cluster: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: outside-cluster: Test not supported in OpenShift\e[0m' WAR: outside-cluster: Test not supported in OpenShift + start_test set-custom-img + '[' 1 -ne 1 ']' + test_name=set-custom-img + echo =========================================================================== =========================================================================== + info 'Rendering files for test set-custom-img' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test set-custom-img\e[0m' Rendering files for test set-custom-img + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p set-custom-img + cd set-custom-img + jaeger_name=my-jaeger + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.image="test"' ./02-install.yaml + '[' true = true ']' + skip_test non-cluster-wide 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=non-cluster-wide + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/set-custom-img + '[' set-custom-img '!=' _build ']' + cd .. + rm -rf non-cluster-wide + warning 'non-cluster-wide: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: non-cluster-wide: Test not supported in OpenShift\e[0m' WAR: non-cluster-wide: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running miscellaneous E2E tests' Running miscellaneous E2E tests + cd tests/e2e/miscellaneous/_build + set +e + KUBECONFIG=/tmp/kubeconfig-488946273 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-rosa-h-ezi9.l97m.s3.devshift.org:443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 8 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/cassandra-spark === PAUSE kuttl/harness/cassandra-spark === RUN kuttl/harness/collector-autoscale === PAUSE kuttl/harness/collector-autoscale === RUN kuttl/harness/collector-otlp-allinone-grpc === PAUSE kuttl/harness/collector-otlp-allinone-grpc === RUN kuttl/harness/collector-otlp-allinone-http === PAUSE kuttl/harness/collector-otlp-allinone-http === RUN kuttl/harness/collector-otlp-production-grpc === PAUSE kuttl/harness/collector-otlp-production-grpc === RUN kuttl/harness/collector-otlp-production-http === PAUSE kuttl/harness/collector-otlp-production-http === RUN kuttl/harness/set-custom-img === PAUSE kuttl/harness/set-custom-img === CONT kuttl/harness/artifacts logger.go:42: 06:57:13 | artifacts | Creating namespace: kuttl-test-actual-starling logger.go:42: 06:57:13 | artifacts | artifacts events from ns kuttl-test-actual-starling: logger.go:42: 06:57:13 | artifacts | Deleting namespace: kuttl-test-actual-starling === CONT kuttl/harness/collector-otlp-allinone-http logger.go:42: 06:57:19 | collector-otlp-allinone-http | Creating namespace: kuttl-test-romantic-dogfish logger.go:42: 06:57:19 | collector-otlp-allinone-http/0-install | starting test step 0-install logger.go:42: 06:57:19 | collector-otlp-allinone-http/0-install | Jaeger:kuttl-test-romantic-dogfish/my-jaeger created logger.go:42: 06:57:51 | collector-otlp-allinone-http/0-install | test step completed 0-install logger.go:42: 06:57:51 | collector-otlp-allinone-http/1-smoke-test | starting test step 1-smoke-test logger.go:42: 06:57:51 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 06:57:54 | collector-otlp-allinone-http/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 06:58:03 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 06:58:04 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 06:58:05 | collector-otlp-allinone-http/1-smoke-test | job.batch/report-span created logger.go:42: 06:58:05 | collector-otlp-allinone-http/1-smoke-test | job.batch/check-span created logger.go:42: 06:58:20 | collector-otlp-allinone-http/1-smoke-test | test step completed 1-smoke-test logger.go:42: 06:58:20 | collector-otlp-allinone-http | collector-otlp-allinone-http events from ns kuttl-test-romantic-dogfish: logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:57:40 +0000 UTC Normal Pod my-jaeger-5688cc8b4-klgh2 Binding Scheduled Successfully assigned kuttl-test-romantic-dogfish/my-jaeger-5688cc8b4-klgh2 to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:57:40 +0000 UTC Normal Pod my-jaeger-5688cc8b4-klgh2 AddedInterface Add eth0 [10.128.0.30/23] from ovn-kubernetes logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:57:40 +0000 UTC Normal Pod my-jaeger-5688cc8b4-klgh2.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" kubelet logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:57:40 +0000 UTC Normal ReplicaSet.apps my-jaeger-5688cc8b4 SuccessfulCreate Created pod: my-jaeger-5688cc8b4-klgh2 replicaset-controller logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:57:40 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-5688cc8b4 to 1 deployment-controller logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:57:45 +0000 UTC Normal Pod my-jaeger-5688cc8b4-klgh2.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" in 4.266s (4.266s including waiting) kubelet logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:57:45 +0000 UTC Normal Pod my-jaeger-5688cc8b4-klgh2.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:57:45 +0000 UTC Normal Pod my-jaeger-5688cc8b4-klgh2.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:57:45 +0000 UTC Normal Pod my-jaeger-5688cc8b4-klgh2.spec.containers{oauth-proxy} Pulling Pulling image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" kubelet logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:57:50 +0000 UTC Normal Pod my-jaeger-5688cc8b4-klgh2.spec.containers{oauth-proxy} Pulled Successfully pulled image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" in 5.756s (5.756s including waiting) kubelet logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:57:51 +0000 UTC Normal Pod my-jaeger-5688cc8b4-klgh2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:57:51 +0000 UTC Normal Pod my-jaeger-5688cc8b4-klgh2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:57:57 +0000 UTC Normal Pod my-jaeger-5688cc8b4-klgh2.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:57:57 +0000 UTC Normal Pod my-jaeger-5688cc8b4-klgh2.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:57:57 +0000 UTC Normal ReplicaSet.apps my-jaeger-5688cc8b4 SuccessfulDelete Deleted pod: my-jaeger-5688cc8b4-klgh2 replicaset-controller logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:57:57 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-5688cc8b4 to 0 from 1 deployment-controller logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:57:58 +0000 UTC Normal Pod my-jaeger-77f58f8f87-549zx Binding Scheduled Successfully assigned kuttl-test-romantic-dogfish/my-jaeger-77f58f8f87-549zx to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:57:58 +0000 UTC Normal ReplicaSet.apps my-jaeger-77f58f8f87 SuccessfulCreate Created pod: my-jaeger-77f58f8f87-549zx replicaset-controller logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:57:58 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-77f58f8f87 to 1 deployment-controller logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:57:59 +0000 UTC Normal Pod my-jaeger-77f58f8f87-549zx AddedInterface Add eth0 [10.128.0.31/23] from ovn-kubernetes logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:57:59 +0000 UTC Normal Pod my-jaeger-77f58f8f87-549zx.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:57:59 +0000 UTC Normal Pod my-jaeger-77f58f8f87-549zx.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:57:59 +0000 UTC Normal Pod my-jaeger-77f58f8f87-549zx.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:57:59 +0000 UTC Normal Pod my-jaeger-77f58f8f87-549zx.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:57:59 +0000 UTC Normal Pod my-jaeger-77f58f8f87-549zx.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:57:59 +0000 UTC Normal Pod my-jaeger-77f58f8f87-549zx.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:58:05 +0000 UTC Normal Pod check-span-rpdxt Binding Scheduled Successfully assigned kuttl-test-romantic-dogfish/check-span-rpdxt to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:58:05 +0000 UTC Normal Pod check-span-rpdxt AddedInterface Add eth0 [10.130.0.28/23] from ovn-kubernetes logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:58:05 +0000 UTC Normal Pod check-span-rpdxt.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:58:05 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-rpdxt job-controller logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:58:05 +0000 UTC Normal Pod report-span-6628b Binding Scheduled Successfully assigned kuttl-test-romantic-dogfish/report-span-6628b to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:58:05 +0000 UTC Normal Pod report-span-6628b AddedInterface Add eth0 [10.129.0.28/23] from ovn-kubernetes logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:58:05 +0000 UTC Normal Pod report-span-6628b.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:58:05 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-6628b job-controller logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:58:08 +0000 UTC Normal Pod check-span-rpdxt.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 3.161s (3.161s including waiting) kubelet logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:58:08 +0000 UTC Normal Pod check-span-rpdxt.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:58:08 +0000 UTC Normal Pod check-span-rpdxt.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:58:08 +0000 UTC Normal Pod report-span-6628b.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 2.869s (2.869s including waiting) kubelet logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:58:08 +0000 UTC Normal Pod report-span-6628b.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:58:08 +0000 UTC Normal Pod report-span-6628b.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 06:58:20 | collector-otlp-allinone-http | 2024-09-26 06:58:19 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 06:58:20 | collector-otlp-allinone-http | Deleting namespace: kuttl-test-romantic-dogfish === CONT kuttl/harness/set-custom-img logger.go:42: 06:58:32 | set-custom-img | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 06:58:32 | set-custom-img | Ignoring check-collector-img.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 06:58:32 | set-custom-img | Creating namespace: kuttl-test-polite-foal logger.go:42: 06:58:32 | set-custom-img/1-install | starting test step 1-install logger.go:42: 06:58:32 | set-custom-img/1-install | Jaeger:kuttl-test-polite-foal/my-jaeger created logger.go:42: 06:59:31 | set-custom-img/1-install | test step completed 1-install logger.go:42: 06:59:31 | set-custom-img/2-install | starting test step 2-install logger.go:42: 06:59:31 | set-custom-img/2-install | Jaeger:kuttl-test-polite-foal/my-jaeger updated logger.go:42: 06:59:31 | set-custom-img/2-install | test step completed 2-install logger.go:42: 06:59:31 | set-custom-img/3-check-image | starting test step 3-check-image logger.go:42: 06:59:31 | set-custom-img/3-check-image | running command: [sh -c ./check-collector-img.sh] logger.go:42: 06:59:32 | set-custom-img/3-check-image | Collector image missmatch. Expected: test. Has: registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c logger.go:42: 06:59:37 | set-custom-img/3-check-image | Collector image asserted properly! logger.go:42: 06:59:37 | set-custom-img/3-check-image | test step completed 3-check-image logger.go:42: 06:59:37 | set-custom-img | set-custom-img events from ns kuttl-test-polite-foal: logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:58:38 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestpolitefoalmyjaeger-1-5564c4df8f SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestpolitefoalmyjaeger-1-5564c4df8fml7z4 replicaset-controller logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:58:38 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpolitefoalmyjaeger-1-5564c4df8fml7z4 Binding Scheduled Successfully assigned kuttl-test-polite-foal/elasticsearch-cdm-kuttltestpolitefoalmyjaeger-1-5564c4df8fml7z4 to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:58:38 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestpolitefoalmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestpolitefoalmyjaeger-1-5564c4df8f to 1 deployment-controller logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:58:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpolitefoalmyjaeger-1-5564c4df8fml7z4 AddedInterface Add eth0 [10.128.0.32/23] from ovn-kubernetes logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:58:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpolitefoalmyjaeger-1-5564c4df8fml7z4.spec.containers{elasticsearch} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:bc97b8e13087050dfb1a75b02c5b780cbb8fb12a849a655169d072ba8bbf42b4" kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:58:48 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpolitefoalmyjaeger-1-5564c4df8fml7z4.spec.containers{elasticsearch} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:bc97b8e13087050dfb1a75b02c5b780cbb8fb12a849a655169d072ba8bbf42b4" in 9.357s (9.357s including waiting) kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:58:48 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpolitefoalmyjaeger-1-5564c4df8fml7z4.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:58:48 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpolitefoalmyjaeger-1-5564c4df8fml7z4.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:58:48 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpolitefoalmyjaeger-1-5564c4df8fml7z4.spec.containers{proxy} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:1d1d7d47b616995d18692f3a2d6232d7e1a0f41bc1503d87100f31939e080a78" kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:58:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpolitefoalmyjaeger-1-5564c4df8fml7z4.spec.containers{proxy} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:1d1d7d47b616995d18692f3a2d6232d7e1a0f41bc1503d87100f31939e080a78" in 6.019s (6.02s including waiting) kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:58:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpolitefoalmyjaeger-1-5564c4df8fml7z4.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:58:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpolitefoalmyjaeger-1-5564c4df8fml7z4.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:58:59 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestpolitefoalmyjaeger-1-5564c4df8fml7z4.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:04 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestpolitefoalmyjaeger-1-5564c4df8fml7z4.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:15 +0000 UTC Normal Pod my-jaeger-collector-545f5796db-dzdf7 Binding Scheduled Successfully assigned kuttl-test-polite-foal/my-jaeger-collector-545f5796db-dzdf7 to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:15 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-545f5796db SuccessfulCreate Created pod: my-jaeger-collector-545f5796db-dzdf7 replicaset-controller logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:15 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-545f5796db to 1 deployment-controller logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:16 +0000 UTC Warning Pod my-jaeger-collector-545f5796db-dzdf7 FailedMount MountVolume.SetUp failed for volume "my-jaeger-collector-tls-config-volume" : secret "my-jaeger-collector-headless-tls" not found kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:16 +0000 UTC Normal Pod my-jaeger-query-6c45996f9-clj5r Binding Scheduled Successfully assigned kuttl-test-polite-foal/my-jaeger-query-6c45996f9-clj5r to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:16 +0000 UTC Normal Pod my-jaeger-query-6c45996f9-clj5r AddedInterface Add eth0 [10.130.0.29/23] from ovn-kubernetes logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:16 +0000 UTC Normal Pod my-jaeger-query-6c45996f9-clj5r.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:16 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-6c45996f9 SuccessfulCreate Created pod: my-jaeger-query-6c45996f9-clj5r replicaset-controller logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:16 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-6c45996f9 to 1 deployment-controller logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:17 +0000 UTC Normal Pod my-jaeger-collector-545f5796db-dzdf7 AddedInterface Add eth0 [10.129.0.29/23] from ovn-kubernetes logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:17 +0000 UTC Normal Pod my-jaeger-collector-545f5796db-dzdf7.spec.containers{jaeger-collector} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:21 +0000 UTC Normal Pod my-jaeger-collector-545f5796db-dzdf7.spec.containers{jaeger-collector} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" in 4.672s (4.672s including waiting) kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:21 +0000 UTC Normal Pod my-jaeger-collector-545f5796db-dzdf7.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:22 +0000 UTC Normal Pod my-jaeger-collector-545f5796db-dzdf7.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:22 +0000 UTC Normal Pod my-jaeger-query-6c45996f9-clj5r.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" in 5.076s (5.076s including waiting) kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:22 +0000 UTC Normal Pod my-jaeger-query-6c45996f9-clj5r.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:22 +0000 UTC Normal Pod my-jaeger-query-6c45996f9-clj5r.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:22 +0000 UTC Normal Pod my-jaeger-query-6c45996f9-clj5r.spec.containers{oauth-proxy} Pulling Pulling image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:25 +0000 UTC Normal Pod my-jaeger-query-6c45996f9-clj5r.spec.containers{oauth-proxy} Pulled Successfully pulled image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" in 3.828s (3.828s including waiting) kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:26 +0000 UTC Normal Pod my-jaeger-query-6c45996f9-clj5r.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:26 +0000 UTC Normal Pod my-jaeger-query-6c45996f9-clj5r.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:26 +0000 UTC Normal Pod my-jaeger-query-6c45996f9-clj5r.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:29 +0000 UTC Normal Pod my-jaeger-query-6c45996f9-clj5r.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" in 3.795s (3.795s including waiting) kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:29 +0000 UTC Normal Pod my-jaeger-query-6c45996f9-clj5r.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:29 +0000 UTC Normal Pod my-jaeger-query-6c45996f9-clj5r.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:32 +0000 UTC Normal Pod my-jaeger-collector-545f5796db-dzdf7.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:32 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-545f5796db SuccessfulDelete Deleted pod: my-jaeger-collector-545f5796db-dzdf7 replicaset-controller logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:32 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-545f5796db to 0 from 1 deployment-controller logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:33 +0000 UTC Normal Pod my-jaeger-collector-844cdff969-g5s6w Binding Scheduled Successfully assigned kuttl-test-polite-foal/my-jaeger-collector-844cdff969-g5s6w to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:33 +0000 UTC Normal Pod my-jaeger-collector-844cdff969-g5s6w AddedInterface Add eth0 [10.129.0.30/23] from ovn-kubernetes logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:33 +0000 UTC Normal Pod my-jaeger-collector-844cdff969-g5s6w.spec.containers{jaeger-collector} Pulling Pulling image "test" kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:33 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-844cdff969 SuccessfulCreate Created pod: my-jaeger-collector-844cdff969-g5s6w replicaset-controller logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:33 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-844cdff969 to 1 deployment-controller logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:34 +0000 UTC Warning Pod my-jaeger-collector-844cdff969-g5s6w.spec.containers{jaeger-collector} Failed Failed to pull image "test": reading manifest latest in docker.io/library/test: requested access to the resource is denied kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:34 +0000 UTC Warning Pod my-jaeger-collector-844cdff969-g5s6w.spec.containers{jaeger-collector} Failed Error: ErrImagePull kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:35 +0000 UTC Normal Pod my-jaeger-collector-844cdff969-g5s6w.spec.containers{jaeger-collector} BackOff Back-off pulling image "test" kubelet logger.go:42: 06:59:37 | set-custom-img | 2024-09-26 06:59:35 +0000 UTC Warning Pod my-jaeger-collector-844cdff969-g5s6w.spec.containers{jaeger-collector} Failed Error: ImagePullBackOff kubelet logger.go:42: 06:59:37 | set-custom-img | Deleting namespace: kuttl-test-polite-foal === CONT kuttl/harness/collector-otlp-production-http logger.go:42: 06:59:43 | collector-otlp-production-http | Creating namespace: kuttl-test-sacred-lobster logger.go:42: 06:59:43 | collector-otlp-production-http/1-install | starting test step 1-install logger.go:42: 06:59:44 | collector-otlp-production-http/1-install | Jaeger:kuttl-test-sacred-lobster/my-jaeger created logger.go:42: 07:00:18 | collector-otlp-production-http/1-install | test step completed 1-install logger.go:42: 07:00:18 | collector-otlp-production-http/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:00:18 | collector-otlp-production-http/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:00:20 | collector-otlp-production-http/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:00:27 | collector-otlp-production-http/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:00:28 | collector-otlp-production-http/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:00:28 | collector-otlp-production-http/2-smoke-test | job.batch/report-span created logger.go:42: 07:00:28 | collector-otlp-production-http/2-smoke-test | job.batch/check-span created logger.go:42: 07:00:42 | collector-otlp-production-http/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:00:42 | collector-otlp-production-http | collector-otlp-production-http events from ns kuttl-test-sacred-lobster: logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 06:59:48 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestsacredlobstermyjaeger-1-77467ddb47 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestsacredlobstermyjaeger-1-77467ddfnjgl replicaset-controller logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 06:59:48 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsacredlobstermyjaeger-1-77467ddfnjgl Binding Scheduled Successfully assigned kuttl-test-sacred-lobster/elasticsearch-cdm-kuttltestsacredlobstermyjaeger-1-77467ddfnjgl to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 06:59:48 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsacredlobstermyjaeger-1-77467ddfnjgl AddedInterface Add eth0 [10.128.0.33/23] from ovn-kubernetes logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 06:59:48 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsacredlobstermyjaeger-1-77467ddfnjgl.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:bc97b8e13087050dfb1a75b02c5b780cbb8fb12a849a655169d072ba8bbf42b4" already present on machine kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 06:59:48 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsacredlobstermyjaeger-1-77467ddfnjgl.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 06:59:48 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsacredlobstermyjaeger-1-77467ddfnjgl.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 06:59:48 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsacredlobstermyjaeger-1-77467ddfnjgl.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:1d1d7d47b616995d18692f3a2d6232d7e1a0f41bc1503d87100f31939e080a78" already present on machine kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 06:59:48 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsacredlobstermyjaeger-1-77467ddfnjgl.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 06:59:48 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsacredlobstermyjaeger-1-77467ddfnjgl.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 06:59:48 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestsacredlobstermyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestsacredlobstermyjaeger-1-77467ddb47 to 1 deployment-controller logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 06:59:58 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestsacredlobstermyjaeger-1-77467ddfnjgl.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:03 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestsacredlobstermyjaeger-1-77467ddfnjgl.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:15 +0000 UTC Normal Pod my-jaeger-collector-bd9676c54-rfcg7 Binding Scheduled Successfully assigned kuttl-test-sacred-lobster/my-jaeger-collector-bd9676c54-rfcg7 to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:15 +0000 UTC Warning Pod my-jaeger-collector-bd9676c54-rfcg7 FailedMount MountVolume.SetUp failed for volume "my-jaeger-collector-tls-config-volume" : secret "my-jaeger-collector-headless-tls" not found kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:15 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-bd9676c54 SuccessfulCreate Created pod: my-jaeger-collector-bd9676c54-rfcg7 replicaset-controller logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:15 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-bd9676c54 to 1 deployment-controller logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:15 +0000 UTC Normal Pod my-jaeger-query-656fffcbc7-qhn8c Binding Scheduled Successfully assigned kuttl-test-sacred-lobster/my-jaeger-query-656fffcbc7-qhn8c to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:15 +0000 UTC Normal Pod my-jaeger-query-656fffcbc7-qhn8c AddedInterface Add eth0 [10.130.0.31/23] from ovn-kubernetes logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:15 +0000 UTC Normal Pod my-jaeger-query-656fffcbc7-qhn8c.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:15 +0000 UTC Normal Pod my-jaeger-query-656fffcbc7-qhn8c.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:15 +0000 UTC Normal Pod my-jaeger-query-656fffcbc7-qhn8c.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:15 +0000 UTC Normal Pod my-jaeger-query-656fffcbc7-qhn8c.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:15 +0000 UTC Normal Pod my-jaeger-query-656fffcbc7-qhn8c.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:15 +0000 UTC Normal Pod my-jaeger-query-656fffcbc7-qhn8c.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:15 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-656fffcbc7 SuccessfulCreate Created pod: my-jaeger-query-656fffcbc7-qhn8c replicaset-controller logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:15 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-656fffcbc7 to 1 deployment-controller logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:16 +0000 UTC Normal Pod my-jaeger-collector-bd9676c54-rfcg7 AddedInterface Add eth0 [10.129.0.31/23] from ovn-kubernetes logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:16 +0000 UTC Normal Pod my-jaeger-collector-bd9676c54-rfcg7.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:16 +0000 UTC Normal Pod my-jaeger-collector-bd9676c54-rfcg7.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:16 +0000 UTC Normal Pod my-jaeger-collector-bd9676c54-rfcg7.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:16 +0000 UTC Normal Pod my-jaeger-query-656fffcbc7-qhn8c.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:16 +0000 UTC Normal Pod my-jaeger-query-656fffcbc7-qhn8c.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:16 +0000 UTC Normal Pod my-jaeger-query-656fffcbc7-qhn8c.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:21 +0000 UTC Normal Pod my-jaeger-query-656fffcbc7-qhn8c.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:21 +0000 UTC Normal Pod my-jaeger-query-656fffcbc7-qhn8c.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:21 +0000 UTC Normal Pod my-jaeger-query-656fffcbc7-qhn8c.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:21 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-656fffcbc7 SuccessfulDelete Deleted pod: my-jaeger-query-656fffcbc7-qhn8c replicaset-controller logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:21 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-656fffcbc7 to 0 from 1 deployment-controller logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:22 +0000 UTC Normal Pod my-jaeger-query-6785856c79-jmbpf Binding Scheduled Successfully assigned kuttl-test-sacred-lobster/my-jaeger-query-6785856c79-jmbpf to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:22 +0000 UTC Normal Pod my-jaeger-query-6785856c79-jmbpf AddedInterface Add eth0 [10.130.0.32/23] from ovn-kubernetes logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:22 +0000 UTC Normal Pod my-jaeger-query-6785856c79-jmbpf.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:22 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-6785856c79 SuccessfulCreate Created pod: my-jaeger-query-6785856c79-jmbpf replicaset-controller logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:22 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-6785856c79 to 1 deployment-controller logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:23 +0000 UTC Normal Pod my-jaeger-query-6785856c79-jmbpf.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:23 +0000 UTC Normal Pod my-jaeger-query-6785856c79-jmbpf.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:23 +0000 UTC Normal Pod my-jaeger-query-6785856c79-jmbpf.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:23 +0000 UTC Normal Pod my-jaeger-query-6785856c79-jmbpf.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:23 +0000 UTC Normal Pod my-jaeger-query-6785856c79-jmbpf.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:23 +0000 UTC Normal Pod my-jaeger-query-6785856c79-jmbpf.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:23 +0000 UTC Normal Pod my-jaeger-query-6785856c79-jmbpf.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:23 +0000 UTC Normal Pod my-jaeger-query-6785856c79-jmbpf.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:28 +0000 UTC Normal Pod check-span-8rjzf Binding Scheduled Successfully assigned kuttl-test-sacred-lobster/check-span-8rjzf to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:28 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-8rjzf job-controller logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:28 +0000 UTC Normal Pod report-span-xf5lk Binding Scheduled Successfully assigned kuttl-test-sacred-lobster/report-span-xf5lk to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:28 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-xf5lk job-controller logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:29 +0000 UTC Normal Pod check-span-8rjzf AddedInterface Add eth0 [10.129.0.33/23] from ovn-kubernetes logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:29 +0000 UTC Normal Pod check-span-8rjzf.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:29 +0000 UTC Normal Pod report-span-xf5lk AddedInterface Add eth0 [10.129.0.32/23] from ovn-kubernetes logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:29 +0000 UTC Normal Pod report-span-xf5lk.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:30 +0000 UTC Normal Pod check-span-8rjzf.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.571s (1.571s including waiting) kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:30 +0000 UTC Normal Pod check-span-8rjzf.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:30 +0000 UTC Normal Pod check-span-8rjzf.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:30 +0000 UTC Normal Pod report-span-xf5lk.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 903ms (903ms including waiting) kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:30 +0000 UTC Normal Pod report-span-xf5lk.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:30 +0000 UTC Normal Pod report-span-xf5lk.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:31 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:31 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:31 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:00:42 | collector-otlp-production-http | 2024-09-26 07:00:42 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:00:43 | collector-otlp-production-http | Deleting namespace: kuttl-test-sacred-lobster === CONT kuttl/harness/collector-otlp-production-grpc logger.go:42: 07:00:49 | collector-otlp-production-grpc | Creating namespace: kuttl-test-crucial-primate logger.go:42: 07:00:49 | collector-otlp-production-grpc/1-install | starting test step 1-install logger.go:42: 07:00:49 | collector-otlp-production-grpc/1-install | Jaeger:kuttl-test-crucial-primate/my-jaeger created logger.go:42: 07:01:23 | collector-otlp-production-grpc/1-install | test step completed 1-install logger.go:42: 07:01:23 | collector-otlp-production-grpc/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:01:23 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:01:24 | collector-otlp-production-grpc/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:01:34 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:01:34 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:01:34 | collector-otlp-production-grpc/2-smoke-test | job.batch/report-span created logger.go:42: 07:01:35 | collector-otlp-production-grpc/2-smoke-test | job.batch/check-span created logger.go:42: 07:01:56 | collector-otlp-production-grpc/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:01:56 | collector-otlp-production-grpc | collector-otlp-production-grpc events from ns kuttl-test-crucial-primate: logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:00:53 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcrucialprimatemyjaeger-1-57754f6vf7z Binding Scheduled Successfully assigned kuttl-test-crucial-primate/elasticsearch-cdm-kuttltestcrucialprimatemyjaeger-1-57754f6vf7z to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:00:53 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcrucialprimatemyjaeger-1-57754f6vf7z AddedInterface Add eth0 [10.128.0.34/23] from ovn-kubernetes logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:00:53 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcrucialprimatemyjaeger-1-57754f6vf7z.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:bc97b8e13087050dfb1a75b02c5b780cbb8fb12a849a655169d072ba8bbf42b4" already present on machine kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:00:53 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcrucialprimatemyjaeger-1-57754f6vf7z.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:00:53 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcrucialprimatemyjaeger-1-57754f6vf7z.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:00:53 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcrucialprimatemyjaeger-1-57754f6vf7z.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:1d1d7d47b616995d18692f3a2d6232d7e1a0f41bc1503d87100f31939e080a78" already present on machine kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:00:53 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcrucialprimatemyjaeger-1-57754f6vf7z.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:00:53 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcrucialprimatemyjaeger-1-57754f6vf7z.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:00:53 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestcrucialprimatemyjaeger-1-57754fcf89 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestcrucialprimatemyjaeger-1-57754f6vf7z replicaset-controller logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:00:53 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestcrucialprimatemyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestcrucialprimatemyjaeger-1-57754fcf89 to 1 deployment-controller logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:03 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestcrucialprimatemyjaeger-1-57754f6vf7z.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:08 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestcrucialprimatemyjaeger-1-57754f6vf7z.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:20 +0000 UTC Normal Pod my-jaeger-collector-5c587fd8dd-6n7pn Binding Scheduled Successfully assigned kuttl-test-crucial-primate/my-jaeger-collector-5c587fd8dd-6n7pn to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:20 +0000 UTC Normal Pod my-jaeger-collector-5c587fd8dd-6n7pn AddedInterface Add eth0 [10.129.0.34/23] from ovn-kubernetes logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:20 +0000 UTC Normal Pod my-jaeger-collector-5c587fd8dd-6n7pn.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:20 +0000 UTC Normal Pod my-jaeger-collector-5c587fd8dd-6n7pn.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:20 +0000 UTC Normal Pod my-jaeger-collector-5c587fd8dd-6n7pn.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:20 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-5c587fd8dd SuccessfulCreate Created pod: my-jaeger-collector-5c587fd8dd-6n7pn replicaset-controller logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:20 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-5c587fd8dd to 1 deployment-controller logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:20 +0000 UTC Normal Pod my-jaeger-query-8874d6ffc-2phpx Binding Scheduled Successfully assigned kuttl-test-crucial-primate/my-jaeger-query-8874d6ffc-2phpx to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:20 +0000 UTC Normal Pod my-jaeger-query-8874d6ffc-2phpx AddedInterface Add eth0 [10.130.0.33/23] from ovn-kubernetes logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:20 +0000 UTC Normal Pod my-jaeger-query-8874d6ffc-2phpx.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:20 +0000 UTC Normal Pod my-jaeger-query-8874d6ffc-2phpx.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:20 +0000 UTC Normal Pod my-jaeger-query-8874d6ffc-2phpx.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:20 +0000 UTC Normal Pod my-jaeger-query-8874d6ffc-2phpx.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:20 +0000 UTC Normal Pod my-jaeger-query-8874d6ffc-2phpx.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:20 +0000 UTC Normal Pod my-jaeger-query-8874d6ffc-2phpx.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:20 +0000 UTC Normal Pod my-jaeger-query-8874d6ffc-2phpx.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:20 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-8874d6ffc SuccessfulCreate Created pod: my-jaeger-query-8874d6ffc-2phpx replicaset-controller logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:20 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-8874d6ffc to 1 deployment-controller logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:21 +0000 UTC Normal Pod my-jaeger-query-8874d6ffc-2phpx.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:21 +0000 UTC Normal Pod my-jaeger-query-8874d6ffc-2phpx.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:27 +0000 UTC Normal Pod my-jaeger-query-8874d6ffc-2phpx.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:27 +0000 UTC Normal Pod my-jaeger-query-8874d6ffc-2phpx.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:27 +0000 UTC Normal Pod my-jaeger-query-8874d6ffc-2phpx.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:27 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-8874d6ffc SuccessfulDelete Deleted pod: my-jaeger-query-8874d6ffc-2phpx replicaset-controller logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:27 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-8874d6ffc to 0 from 1 deployment-controller logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:28 +0000 UTC Normal Pod my-jaeger-query-78c5b6bf5c-grs8w Binding Scheduled Successfully assigned kuttl-test-crucial-primate/my-jaeger-query-78c5b6bf5c-grs8w to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:28 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-78c5b6bf5c SuccessfulCreate Created pod: my-jaeger-query-78c5b6bf5c-grs8w replicaset-controller logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:28 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-78c5b6bf5c to 1 deployment-controller logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:29 +0000 UTC Normal Pod my-jaeger-query-78c5b6bf5c-grs8w AddedInterface Add eth0 [10.130.0.34/23] from ovn-kubernetes logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:29 +0000 UTC Normal Pod my-jaeger-query-78c5b6bf5c-grs8w.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:29 +0000 UTC Normal Pod my-jaeger-query-78c5b6bf5c-grs8w.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:29 +0000 UTC Normal Pod my-jaeger-query-78c5b6bf5c-grs8w.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:29 +0000 UTC Normal Pod my-jaeger-query-78c5b6bf5c-grs8w.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:29 +0000 UTC Normal Pod my-jaeger-query-78c5b6bf5c-grs8w.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:29 +0000 UTC Normal Pod my-jaeger-query-78c5b6bf5c-grs8w.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:29 +0000 UTC Normal Pod my-jaeger-query-78c5b6bf5c-grs8w.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:29 +0000 UTC Normal Pod my-jaeger-query-78c5b6bf5c-grs8w.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:29 +0000 UTC Normal Pod my-jaeger-query-78c5b6bf5c-grs8w.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:34 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-8fc7k job-controller logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:35 +0000 UTC Normal Pod check-span-n2x98 Binding Scheduled Successfully assigned kuttl-test-crucial-primate/check-span-n2x98 to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:35 +0000 UTC Normal Pod check-span-n2x98 AddedInterface Add eth0 [10.129.0.36/23] from ovn-kubernetes logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:35 +0000 UTC Normal Pod check-span-n2x98.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:35 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-n2x98 job-controller logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:35 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:35 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:35 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:35 +0000 UTC Normal Pod report-span-8fc7k Binding Scheduled Successfully assigned kuttl-test-crucial-primate/report-span-8fc7k to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:35 +0000 UTC Normal Pod report-span-8fc7k AddedInterface Add eth0 [10.129.0.35/23] from ovn-kubernetes logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:35 +0000 UTC Normal Pod report-span-8fc7k.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:36 +0000 UTC Normal Pod check-span-n2x98.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 934ms (934ms including waiting) kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:36 +0000 UTC Normal Pod check-span-n2x98.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:36 +0000 UTC Normal Pod check-span-n2x98.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:36 +0000 UTC Normal Pod report-span-8fc7k.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 816ms (816ms including waiting) kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:36 +0000 UTC Normal Pod report-span-8fc7k.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:36 +0000 UTC Normal Pod report-span-8fc7k.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:01:56 | collector-otlp-production-grpc | 2024-09-26 07:01:55 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:01:56 | collector-otlp-production-grpc | Deleting namespace: kuttl-test-crucial-primate === CONT kuttl/harness/collector-autoscale logger.go:42: 07:02:08 | collector-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:02:08 | collector-autoscale | Creating namespace: kuttl-test-famous-hare logger.go:42: 07:02:08 | collector-autoscale/1-install | starting test step 1-install logger.go:42: 07:02:09 | collector-autoscale/1-install | Jaeger:kuttl-test-famous-hare/simple-prod created logger.go:42: 07:02:44 | collector-autoscale/1-install | test step completed 1-install logger.go:42: 07:02:44 | collector-autoscale/2- | starting test step 2- logger.go:42: 07:02:45 | collector-autoscale/2- | test step completed 2- logger.go:42: 07:02:45 | collector-autoscale | collector-autoscale events from ns kuttl-test-famous-hare: logger.go:42: 07:02:45 | collector-autoscale | 2024-09-26 07:02:14 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestfamousharesimpleprod-1-5b86bb5bf5 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestfamousharesimpleprod-1-5b86bb5bkd9fc replicaset-controller logger.go:42: 07:02:45 | collector-autoscale | 2024-09-26 07:02:14 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfamousharesimpleprod-1-5b86bb5bkd9fc Binding Scheduled Successfully assigned kuttl-test-famous-hare/elasticsearch-cdm-kuttltestfamousharesimpleprod-1-5b86bb5bkd9fc to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:02:45 | collector-autoscale | 2024-09-26 07:02:14 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfamousharesimpleprod-1-5b86bb5bkd9fc AddedInterface Add eth0 [10.128.0.35/23] from ovn-kubernetes logger.go:42: 07:02:45 | collector-autoscale | 2024-09-26 07:02:14 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestfamousharesimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestfamousharesimpleprod-1-5b86bb5bf5 to 1 deployment-controller logger.go:42: 07:02:45 | collector-autoscale | 2024-09-26 07:02:15 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfamousharesimpleprod-1-5b86bb5bkd9fc.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:bc97b8e13087050dfb1a75b02c5b780cbb8fb12a849a655169d072ba8bbf42b4" already present on machine kubelet logger.go:42: 07:02:45 | collector-autoscale | 2024-09-26 07:02:15 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfamousharesimpleprod-1-5b86bb5bkd9fc.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:02:45 | collector-autoscale | 2024-09-26 07:02:15 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfamousharesimpleprod-1-5b86bb5bkd9fc.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:02:45 | collector-autoscale | 2024-09-26 07:02:15 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfamousharesimpleprod-1-5b86bb5bkd9fc.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:1d1d7d47b616995d18692f3a2d6232d7e1a0f41bc1503d87100f31939e080a78" already present on machine kubelet logger.go:42: 07:02:45 | collector-autoscale | 2024-09-26 07:02:15 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfamousharesimpleprod-1-5b86bb5bkd9fc.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:02:45 | collector-autoscale | 2024-09-26 07:02:15 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfamousharesimpleprod-1-5b86bb5bkd9fc.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:02:45 | collector-autoscale | 2024-09-26 07:02:30 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestfamousharesimpleprod-1-5b86bb5bkd9fc.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:02:45 | collector-autoscale | 2024-09-26 07:02:41 +0000 UTC Normal Pod simple-prod-collector-fcbc49f96-p5nzg Binding Scheduled Successfully assigned kuttl-test-famous-hare/simple-prod-collector-fcbc49f96-p5nzg to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:02:45 | collector-autoscale | 2024-09-26 07:02:41 +0000 UTC Normal Pod simple-prod-collector-fcbc49f96-p5nzg AddedInterface Add eth0 [10.129.0.37/23] from ovn-kubernetes logger.go:42: 07:02:45 | collector-autoscale | 2024-09-26 07:02:41 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-fcbc49f96 SuccessfulCreate Created pod: simple-prod-collector-fcbc49f96-p5nzg replicaset-controller logger.go:42: 07:02:45 | collector-autoscale | 2024-09-26 07:02:41 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-fcbc49f96 to 1 deployment-controller logger.go:42: 07:02:45 | collector-autoscale | 2024-09-26 07:02:41 +0000 UTC Normal Pod simple-prod-query-68d59b55b5-jbwrz Binding Scheduled Successfully assigned kuttl-test-famous-hare/simple-prod-query-68d59b55b5-jbwrz to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:02:45 | collector-autoscale | 2024-09-26 07:02:41 +0000 UTC Normal ReplicaSet.apps simple-prod-query-68d59b55b5 SuccessfulCreate Created pod: simple-prod-query-68d59b55b5-jbwrz replicaset-controller logger.go:42: 07:02:45 | collector-autoscale | 2024-09-26 07:02:41 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-68d59b55b5 to 1 deployment-controller logger.go:42: 07:02:45 | collector-autoscale | 2024-09-26 07:02:42 +0000 UTC Normal Pod simple-prod-collector-fcbc49f96-p5nzg.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 07:02:45 | collector-autoscale | 2024-09-26 07:02:42 +0000 UTC Normal Pod simple-prod-collector-fcbc49f96-p5nzg.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:02:45 | collector-autoscale | 2024-09-26 07:02:42 +0000 UTC Normal Pod simple-prod-collector-fcbc49f96-p5nzg.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:02:45 | collector-autoscale | 2024-09-26 07:02:42 +0000 UTC Normal Pod simple-prod-query-68d59b55b5-jbwrz AddedInterface Add eth0 [10.130.0.35/23] from ovn-kubernetes logger.go:42: 07:02:45 | collector-autoscale | 2024-09-26 07:02:42 +0000 UTC Normal Pod simple-prod-query-68d59b55b5-jbwrz.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:02:45 | collector-autoscale | 2024-09-26 07:02:42 +0000 UTC Normal Pod simple-prod-query-68d59b55b5-jbwrz.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:02:45 | collector-autoscale | 2024-09-26 07:02:42 +0000 UTC Normal Pod simple-prod-query-68d59b55b5-jbwrz.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:02:45 | collector-autoscale | 2024-09-26 07:02:42 +0000 UTC Normal Pod simple-prod-query-68d59b55b5-jbwrz.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:02:45 | collector-autoscale | 2024-09-26 07:02:42 +0000 UTC Normal Pod simple-prod-query-68d59b55b5-jbwrz.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:02:45 | collector-autoscale | 2024-09-26 07:02:42 +0000 UTC Normal Pod simple-prod-query-68d59b55b5-jbwrz.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:02:45 | collector-autoscale | 2024-09-26 07:02:42 +0000 UTC Normal Pod simple-prod-query-68d59b55b5-jbwrz.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:02:45 | collector-autoscale | 2024-09-26 07:02:42 +0000 UTC Normal Pod simple-prod-query-68d59b55b5-jbwrz.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:02:45 | collector-autoscale | 2024-09-26 07:02:42 +0000 UTC Normal Pod simple-prod-query-68d59b55b5-jbwrz.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:02:45 | collector-autoscale | Deleting namespace: kuttl-test-famous-hare === CONT kuttl/harness/collector-otlp-allinone-grpc logger.go:42: 07:02:51 | collector-otlp-allinone-grpc | Creating namespace: kuttl-test-leading-polliwog logger.go:42: 07:02:51 | collector-otlp-allinone-grpc/0-install | starting test step 0-install logger.go:42: 07:02:51 | collector-otlp-allinone-grpc/0-install | Jaeger:kuttl-test-leading-polliwog/my-jaeger created logger.go:42: 07:02:59 | collector-otlp-allinone-grpc/0-install | test step completed 0-install logger.go:42: 07:02:59 | collector-otlp-allinone-grpc/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:02:59 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:03:01 | collector-otlp-allinone-grpc/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:03:08 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:03:09 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:03:09 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/report-span created logger.go:42: 07:03:09 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/check-span created logger.go:42: 07:03:31 | collector-otlp-allinone-grpc/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | collector-otlp-allinone-grpc events from ns kuttl-test-leading-polliwog: logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:02:52 +0000 UTC Normal Pod my-jaeger-79fbc6b9d5-pkggq Binding Scheduled Successfully assigned kuttl-test-leading-polliwog/my-jaeger-79fbc6b9d5-pkggq to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:02:52 +0000 UTC Normal Pod my-jaeger-79fbc6b9d5-pkggq AddedInterface Add eth0 [10.130.0.36/23] from ovn-kubernetes logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:02:52 +0000 UTC Normal Pod my-jaeger-79fbc6b9d5-pkggq.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" kubelet logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:02:52 +0000 UTC Normal ReplicaSet.apps my-jaeger-79fbc6b9d5 SuccessfulCreate Created pod: my-jaeger-79fbc6b9d5-pkggq replicaset-controller logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:02:52 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-79fbc6b9d5 to 1 deployment-controller logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:02:57 +0000 UTC Normal Pod my-jaeger-79fbc6b9d5-pkggq.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" in 4.465s (4.465s including waiting) kubelet logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:02:57 +0000 UTC Normal Pod my-jaeger-79fbc6b9d5-pkggq.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:02:57 +0000 UTC Normal Pod my-jaeger-79fbc6b9d5-pkggq.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:02:57 +0000 UTC Normal Pod my-jaeger-79fbc6b9d5-pkggq.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:02:57 +0000 UTC Normal Pod my-jaeger-79fbc6b9d5-pkggq.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:02:57 +0000 UTC Normal Pod my-jaeger-79fbc6b9d5-pkggq.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:03:01 +0000 UTC Normal Pod my-jaeger-79fbc6b9d5-pkggq.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:03:01 +0000 UTC Normal Pod my-jaeger-79fbc6b9d5-pkggq.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:03:01 +0000 UTC Normal ReplicaSet.apps my-jaeger-79fbc6b9d5 SuccessfulDelete Deleted pod: my-jaeger-79fbc6b9d5-pkggq replicaset-controller logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:03:01 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-79fbc6b9d5 to 0 from 1 deployment-controller logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:03:02 +0000 UTC Normal Pod my-jaeger-7764455984-5xnq4 Binding Scheduled Successfully assigned kuttl-test-leading-polliwog/my-jaeger-7764455984-5xnq4 to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:03:02 +0000 UTC Normal ReplicaSet.apps my-jaeger-7764455984 SuccessfulCreate Created pod: my-jaeger-7764455984-5xnq4 replicaset-controller logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:03:02 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-7764455984 to 1 deployment-controller logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:03:03 +0000 UTC Normal Pod my-jaeger-7764455984-5xnq4 AddedInterface Add eth0 [10.130.0.37/23] from ovn-kubernetes logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:03:03 +0000 UTC Normal Pod my-jaeger-7764455984-5xnq4.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:03:03 +0000 UTC Normal Pod my-jaeger-7764455984-5xnq4.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:03:03 +0000 UTC Normal Pod my-jaeger-7764455984-5xnq4.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:03:03 +0000 UTC Normal Pod my-jaeger-7764455984-5xnq4.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:03:03 +0000 UTC Normal Pod my-jaeger-7764455984-5xnq4.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:03:03 +0000 UTC Normal Pod my-jaeger-7764455984-5xnq4.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:03:09 +0000 UTC Normal Pod check-span-rmj9c Binding Scheduled Successfully assigned kuttl-test-leading-polliwog/check-span-rmj9c to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:03:09 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-rmj9c job-controller logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:03:09 +0000 UTC Normal Pod report-span-2jc8h Binding Scheduled Successfully assigned kuttl-test-leading-polliwog/report-span-2jc8h to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:03:09 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-2jc8h job-controller logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:03:10 +0000 UTC Normal Pod check-span-rmj9c AddedInterface Add eth0 [10.129.0.38/23] from ovn-kubernetes logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:03:10 +0000 UTC Normal Pod check-span-rmj9c.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:03:10 +0000 UTC Normal Pod report-span-2jc8h AddedInterface Add eth0 [10.128.0.36/23] from ovn-kubernetes logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:03:10 +0000 UTC Normal Pod report-span-2jc8h.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:03:11 +0000 UTC Normal Pod check-span-rmj9c.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 961ms (961ms including waiting) kubelet logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:03:11 +0000 UTC Normal Pod check-span-rmj9c.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:03:11 +0000 UTC Normal Pod check-span-rmj9c.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:03:13 +0000 UTC Normal Pod report-span-2jc8h.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 3.088s (3.088s including waiting) kubelet logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:03:13 +0000 UTC Normal Pod report-span-2jc8h.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:03:13 +0000 UTC Normal Pod report-span-2jc8h.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | 2024-09-26 07:03:30 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:03:31 | collector-otlp-allinone-grpc | Deleting namespace: kuttl-test-leading-polliwog === CONT kuttl/harness/cassandra-spark logger.go:42: 07:03:43 | cassandra-spark | Ignoring 01-assert.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:03:43 | cassandra-spark | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:03:43 | cassandra-spark | Creating namespace: kuttl-test-welcomed-doberman logger.go:42: 07:03:43 | cassandra-spark | cassandra-spark events from ns kuttl-test-welcomed-doberman: logger.go:42: 07:03:43 | cassandra-spark | Deleting namespace: kuttl-test-welcomed-doberman === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (396.41s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.98s) --- PASS: kuttl/harness/collector-otlp-allinone-http (73.01s) --- PASS: kuttl/harness/set-custom-img (71.75s) --- PASS: kuttl/harness/collector-otlp-production-http (65.49s) --- PASS: kuttl/harness/collector-otlp-production-grpc (79.44s) --- PASS: kuttl/harness/collector-autoscale (42.53s) --- PASS: kuttl/harness/collector-otlp-allinone-grpc (51.74s) --- PASS: kuttl/harness/cassandra-spark (5.89s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name miscellaneous --report --output /logs/artifacts/miscellaneous.xml ./artifacts/kuttl-report.xml time="2024-09-26T07:03:50Z" level=debug msg="Setting a new name for the test suites" time="2024-09-26T07:03:50Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-09-26T07:03:50Z" level=debug msg="normalizing test case names" time="2024-09-26T07:03:50Z" level=debug msg="miscellaneous/artifacts -> miscellaneous_artifacts" time="2024-09-26T07:03:50Z" level=debug msg="miscellaneous/collector-otlp-allinone-http -> miscellaneous_collector_otlp_allinone_http" time="2024-09-26T07:03:50Z" level=debug msg="miscellaneous/set-custom-img -> miscellaneous_set_custom_img" time="2024-09-26T07:03:50Z" level=debug msg="miscellaneous/collector-otlp-production-http -> miscellaneous_collector_otlp_production_http" time="2024-09-26T07:03:50Z" level=debug msg="miscellaneous/collector-otlp-production-grpc -> miscellaneous_collector_otlp_production_grpc" time="2024-09-26T07:03:50Z" level=debug msg="miscellaneous/collector-autoscale -> miscellaneous_collector_autoscale" time="2024-09-26T07:03:50Z" level=debug msg="miscellaneous/collector-otlp-allinone-grpc -> miscellaneous_collector_otlp_allinone_grpc" time="2024-09-26T07:03:50Z" level=debug msg="miscellaneous/cassandra-spark -> miscellaneous_cassandra_spark" +----------------------------------------------+--------+ | NAME | RESULT | +----------------------------------------------+--------+ | miscellaneous_artifacts | passed | | miscellaneous_collector_otlp_allinone_http | passed | | miscellaneous_set_custom_img | passed | | miscellaneous_collector_otlp_production_http | passed | | miscellaneous_collector_otlp_production_grpc | passed | | miscellaneous_collector_autoscale | passed | | miscellaneous_collector_otlp_allinone_grpc | passed | | miscellaneous_cassandra_spark | passed | +----------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh elasticsearch false true + '[' 3 -ne 3 ']' + test_suite_name=elasticsearch + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/elasticsearch.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-elasticsearch make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true \ KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ ./tests/e2e/elasticsearch/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.32 True False 17m Cluster version is 4.15.32' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.32 True False 17m Cluster version is 4.15.32' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/elasticsearch/render.sh ++ export SUITE_DIR=./tests/e2e/elasticsearch ++ SUITE_DIR=./tests/e2e/elasticsearch ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/elasticsearch ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + start_test es-from-aio-to-production + '[' 1 -ne 1 ']' + test_name=es-from-aio-to-production + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-from-aio-to-production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-from-aio-to-production\e[0m' Rendering files for test es-from-aio-to-production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-from-aio-to-production + cd es-from-aio-to-production + jaeger_name=my-jaeger + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 03 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=03 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./03-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch.redundancyPolicy="ZeroRedundancy"' ./03-install.yaml + render_smoke_test my-jaeger true 04 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=04 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./04-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./04-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test es-increasing-replicas + '[' 1 -ne 1 ']' + test_name=es-increasing-replicas + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-increasing-replicas' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-increasing-replicas\e[0m' Rendering files for test es-increasing-replicas + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-from-aio-to-production + '[' es-from-aio-to-production '!=' _build ']' + cd .. + mkdir -p es-increasing-replicas + cd es-increasing-replicas + jaeger_name=simple-prod + '[' true = true ']' + jaeger_deployment_mode=production_autoprovisioned + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.replicas=2 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.query.replicas=2 ./02-install.yaml + cp ./01-assert.yaml ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.replicas=2 ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .status.readyReplicas=2 ./02-assert.yaml + render_smoke_test simple-prod true 03 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=03 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./03-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + cp ./02-install.yaml ./04-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.elasticsearch.nodeCount=2 ./04-install.yaml + /tmp/jaeger-tests/bin/gomplate -f ./openshift-check-es-nodes.yaml.template -o ./05-check-es-nodes.yaml + '[' true = true ']' + skip_test es-index-cleaner-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-increasing-replicas + '[' es-increasing-replicas '!=' _build ']' + cd .. + rm -rf es-index-cleaner-upstream + warning 'es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_index_cleaner -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-index-cleaner-autoprov + '[' 1 -ne 1 ']' + test_name=es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-index-cleaner-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-index-cleaner-autoprov\e[0m' Rendering files for test es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-index-cleaner-autoprov + cd es-index-cleaner-autoprov + jaeger_name=test-es-index-cleaner-with-prefix + cronjob_name=test-es-index-cleaner-with-prefix-es-index-cleaner + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + cp ../../es-index-cleaner-upstream/04-assert.yaml ../../es-index-cleaner-upstream/README.md . + render_install_jaeger test-es-index-cleaner-with-prefix production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options.es.index-prefix=""' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.enabled=false ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.numberOfDays=0 ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.esIndexCleaner.schedule="*/1 * * * *"' ./01-install.yaml + render_report_spans test-es-index-cleaner-with-prefix true 5 00 true 02 + '[' 6 -ne 6 ']' + jaeger=test-es-index-cleaner-with-prefix + is_secured=true + number_of_spans=5 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + export JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=5 + DAYS=5 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + sed 's~enabled: false~enabled: true~gi' ./01-install.yaml + CRONJOB_NAME=test-es-index-cleaner-with-prefix-es-index-cleaner + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./04-wait-es-index-cleaner.yaml + /tmp/jaeger-tests/bin/gomplate -f ./01-install.yaml -o ./05-install.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 00 06 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=00 + test_step=06 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=test-es-index-cleaner-with-prefix-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=test-es-index-cleaner-with-prefix-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./06-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./06-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.16},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.12"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.16},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.12"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.16},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.12"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.12 ++ version_ge 5.8.12 5.4 +++ echo 5.8.12 5.4 +++ tr ' ' '\n' +++ sort -rV +++ head -n 1 ++ test 5.8.12 == 5.8.12 + '[' -n '' ']' + skip_test es-index-cleaner-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-index-cleaner-autoprov + '[' es-index-cleaner-autoprov '!=' _build ']' + cd .. + rm -rf es-index-cleaner-managed + warning 'es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + start_test es-multiinstance + '[' 1 -ne 1 ']' + test_name=es-multiinstance + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-multiinstance' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-multiinstance\e[0m' Rendering files for test es-multiinstance + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-multiinstance + cd es-multiinstance + jaeger_name=instance-1 + render_install_jaeger instance-1 production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=instance-1 + JAEGER_NAME=instance-1 + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f ./03-create-second-instance.yaml.template -o 03-create-second-instance.yaml + '[' true = true ']' + skip_test es-rollover-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-rollover-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-multiinstance + '[' es-multiinstance '!=' _build ']' + cd .. + rm -rf es-rollover-upstream + warning 'es-rollover-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_rollover -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-rollover-autoprov + '[' 1 -ne 1 ']' + test_name=es-rollover-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-rollover-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-rollover-autoprov\e[0m' Rendering files for test es-rollover-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-rollover-autoprov + cd es-rollover-autoprov + cp ../../es-rollover-upstream/05-assert.yaml ../../es-rollover-upstream/05-install.yaml ../../es-rollover-upstream/README.md . + jaeger_name=my-jaeger + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_report_spans my-jaeger true 2 00 true 02 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 00 03 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=00 + test_step=03 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./03-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./03-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 01 04 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=01 + test_step=04 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=01 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./04-check-indices.yaml + JOB_NUMBER=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./04-assert.yaml + render_report_spans my-jaeger true 2 02 true 06 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=02 + ensure_reported_spans=true + test_step=06 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=02 + JOB_NUMBER=02 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./06-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./06-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 02 07 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=02 + test_step=07 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=02 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./07-check-indices.yaml + JOB_NUMBER=02 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./07-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' 03 08 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + job_number=03 + test_step=08 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=03 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./08-check-indices.yaml + JOB_NUMBER=03 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./08-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' 04 09 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + job_number=04 + test_step=09 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=04 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./09-check-indices.yaml + JOB_NUMBER=04 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./09-assert.yaml + render_report_spans my-jaeger true 2 03 true 10 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=03 + ensure_reported_spans=true + test_step=10 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=03 + JOB_NUMBER=03 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./10-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./10-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + CRONJOB_NAME=my-jaeger-es-rollover + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./11-wait-rollover.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-000002'\'',' 05 11 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-000002'\'',' + job_number=05 + test_step=11 + escape_command ''\''--name'\'', '\''jaeger-span-000002'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-000002'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-000002'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-000002'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=05 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./11-check-indices.yaml + JOB_NUMBER=05 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./11-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' 06 12 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + job_number=06 + test_step=12 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=06 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./12-check-indices.yaml + JOB_NUMBER=06 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./12-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.16},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.12"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.16},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.12"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.16},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.12"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.12 ++ version_ge 5.8.12 5.4 +++ echo 5.8.12 5.4 +++ tr ' ' '\n' +++ sort -rV +++ head -n 1 ++ test 5.8.12 == 5.8.12 + '[' -n '' ']' + skip_test es-rollover-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-rollover-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-rollover-autoprov + '[' es-rollover-autoprov '!=' _build ']' + cd .. + rm -rf es-rollover-managed + warning 'es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + skip_test es-spark-dependencies 'This test is not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=es-spark-dependencies + message='This test is not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + rm -rf es-spark-dependencies + warning 'es-spark-dependencies: This test is not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-spark-dependencies: This test is not supported in OpenShift\e[0m' WAR: es-spark-dependencies: This test is not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running elasticsearch E2E tests' Running elasticsearch E2E tests + cd tests/e2e/elasticsearch/_build + set +e + KUBECONFIG=/tmp/kubeconfig-488946273 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-rosa-h-ezi9.l97m.s3.devshift.org:443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 7 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/es-from-aio-to-production === PAUSE kuttl/harness/es-from-aio-to-production === RUN kuttl/harness/es-increasing-replicas === PAUSE kuttl/harness/es-increasing-replicas === RUN kuttl/harness/es-index-cleaner-autoprov === PAUSE kuttl/harness/es-index-cleaner-autoprov === RUN kuttl/harness/es-multiinstance === PAUSE kuttl/harness/es-multiinstance === RUN kuttl/harness/es-rollover-autoprov === PAUSE kuttl/harness/es-rollover-autoprov === RUN kuttl/harness/es-simple-prod === PAUSE kuttl/harness/es-simple-prod === CONT kuttl/harness/artifacts logger.go:42: 07:04:16 | artifacts | Creating namespace: kuttl-test-golden-longhorn logger.go:42: 07:04:16 | artifacts | artifacts events from ns kuttl-test-golden-longhorn: logger.go:42: 07:04:16 | artifacts | Deleting namespace: kuttl-test-golden-longhorn === CONT kuttl/harness/es-multiinstance logger.go:42: 07:04:22 | es-multiinstance | Ignoring 03-create-second-instance.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:04:22 | es-multiinstance | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:04:22 | es-multiinstance | Creating namespace: kuttl-test-exciting-ghoul logger.go:42: 07:04:22 | es-multiinstance/0-clear-namespace | starting test step 0-clear-namespace logger.go:42: 07:04:22 | es-multiinstance/0-clear-namespace | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --ignore-not-found=true] logger.go:42: 07:04:23 | es-multiinstance/0-clear-namespace | test step completed 0-clear-namespace logger.go:42: 07:04:23 | es-multiinstance/1-install | starting test step 1-install logger.go:42: 07:04:23 | es-multiinstance/1-install | Jaeger:kuttl-test-exciting-ghoul/instance-1 created logger.go:42: 07:04:59 | es-multiinstance/1-install | test step completed 1-install logger.go:42: 07:04:59 | es-multiinstance/2-create-namespace | starting test step 2-create-namespace logger.go:42: 07:04:59 | es-multiinstance/2-create-namespace | running command: [sh -c kubectl create namespace jaeger-e2e-multiinstance-test] logger.go:42: 07:04:59 | es-multiinstance/2-create-namespace | namespace/jaeger-e2e-multiinstance-test created logger.go:42: 07:04:59 | es-multiinstance/2-create-namespace | test step completed 2-create-namespace logger.go:42: 07:04:59 | es-multiinstance/3-create-second-instance | starting test step 3-create-second-instance logger.go:42: 07:04:59 | es-multiinstance/3-create-second-instance | running command: [sh -c kubectl apply -f ./01-install.yaml -n jaeger-e2e-multiinstance-test] logger.go:42: 07:05:00 | es-multiinstance/3-create-second-instance | jaeger.jaegertracing.io/instance-1 created logger.go:42: 07:05:00 | es-multiinstance/3-create-second-instance | running command: [sh -c /tmp/jaeger-tests/bin/kubectl-kuttl assert ./01-assert.yaml -n jaeger-e2e-multiinstance-test --timeout 1000] logger.go:42: 07:05:35 | es-multiinstance/3-create-second-instance | assert is valid logger.go:42: 07:05:36 | es-multiinstance/3-create-second-instance | test step completed 3-create-second-instance logger.go:42: 07:05:36 | es-multiinstance/4-check-secrets | starting test step 4-check-secrets logger.go:42: 07:05:36 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n $NAMESPACE > secret1] logger.go:42: 07:05:36 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n jaeger-e2e-multiinstance-test > secret2] logger.go:42: 07:05:36 | es-multiinstance/4-check-secrets | running command: [sh -c cmp --silent secret1 secret2 || exit 0] logger.go:42: 07:05:36 | es-multiinstance/4-check-secrets | test step completed 4-check-secrets logger.go:42: 07:05:36 | es-multiinstance/5-delete | starting test step 5-delete logger.go:42: 07:05:36 | es-multiinstance/5-delete | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --wait=false] logger.go:42: 07:05:37 | es-multiinstance/5-delete | namespace "jaeger-e2e-multiinstance-test" deleted logger.go:42: 07:05:37 | es-multiinstance/5-delete | test step completed 5-delete logger.go:42: 07:05:37 | es-multiinstance | es-multiinstance events from ns kuttl-test-exciting-ghoul: logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:29 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestexcitingghoulinstance1-1-59f9b567d7 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestexcitingghoulinstance1-1-59f9b569zsq replicaset-controller logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestexcitingghoulinstance1-1-59f9b569zsq Binding Scheduled Successfully assigned kuttl-test-exciting-ghoul/elasticsearch-cdm-kuttltestexcitingghoulinstance1-1-59f9b569zsq to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestexcitingghoulinstance1-1-59f9b569zsq AddedInterface Add eth0 [10.128.0.37/23] from ovn-kubernetes logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestexcitingghoulinstance1-1-59f9b569zsq.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:bc97b8e13087050dfb1a75b02c5b780cbb8fb12a849a655169d072ba8bbf42b4" already present on machine kubelet logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestexcitingghoulinstance1-1-59f9b569zsq.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestexcitingghoulinstance1-1-59f9b569zsq.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestexcitingghoulinstance1-1-59f9b569zsq.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:1d1d7d47b616995d18692f3a2d6232d7e1a0f41bc1503d87100f31939e080a78" already present on machine kubelet logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestexcitingghoulinstance1-1-59f9b569zsq.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestexcitingghoulinstance1-1-59f9b569zsq.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:29 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestexcitingghoulinstance1-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestexcitingghoulinstance1-1-59f9b567d7 to 1 deployment-controller logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:39 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestexcitingghoulinstance1-1-59f9b569zsq.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:44 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestexcitingghoulinstance1-1-59f9b569zsq.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:56 +0000 UTC Normal Pod instance-1-collector-659f9f9c9f-dxbv2 Binding Scheduled Successfully assigned kuttl-test-exciting-ghoul/instance-1-collector-659f9f9c9f-dxbv2 to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:56 +0000 UTC Normal Pod instance-1-collector-659f9f9c9f-dxbv2 AddedInterface Add eth0 [10.129.0.39/23] from ovn-kubernetes logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:56 +0000 UTC Normal Pod instance-1-collector-659f9f9c9f-dxbv2.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:56 +0000 UTC Normal Pod instance-1-collector-659f9f9c9f-dxbv2.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:56 +0000 UTC Normal Pod instance-1-collector-659f9f9c9f-dxbv2.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:56 +0000 UTC Normal ReplicaSet.apps instance-1-collector-659f9f9c9f SuccessfulCreate Created pod: instance-1-collector-659f9f9c9f-dxbv2 replicaset-controller logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:56 +0000 UTC Normal Deployment.apps instance-1-collector ScalingReplicaSet Scaled up replica set instance-1-collector-659f9f9c9f to 1 deployment-controller logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:56 +0000 UTC Normal Pod instance-1-query-79df9c7dc6-64zjl Binding Scheduled Successfully assigned kuttl-test-exciting-ghoul/instance-1-query-79df9c7dc6-64zjl to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:56 +0000 UTC Normal Pod instance-1-query-79df9c7dc6-64zjl AddedInterface Add eth0 [10.130.0.38/23] from ovn-kubernetes logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:56 +0000 UTC Normal Pod instance-1-query-79df9c7dc6-64zjl.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:56 +0000 UTC Normal Pod instance-1-query-79df9c7dc6-64zjl.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:56 +0000 UTC Normal Pod instance-1-query-79df9c7dc6-64zjl.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:56 +0000 UTC Normal Pod instance-1-query-79df9c7dc6-64zjl.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:56 +0000 UTC Normal Pod instance-1-query-79df9c7dc6-64zjl.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:56 +0000 UTC Normal Pod instance-1-query-79df9c7dc6-64zjl.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:56 +0000 UTC Normal Pod instance-1-query-79df9c7dc6-64zjl.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:56 +0000 UTC Normal Pod instance-1-query-79df9c7dc6-64zjl.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:56 +0000 UTC Normal Pod instance-1-query-79df9c7dc6-64zjl.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:56 +0000 UTC Normal ReplicaSet.apps instance-1-query-79df9c7dc6 SuccessfulCreate Created pod: instance-1-query-79df9c7dc6-64zjl replicaset-controller logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:04:56 +0000 UTC Normal Deployment.apps instance-1-query ScalingReplicaSet Scaled up replica set instance-1-query-79df9c7dc6 to 1 deployment-controller logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:05:11 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:05:11 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:05:37 | es-multiinstance | 2024-09-26 07:05:11 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:05:37 | es-multiinstance | Deleting namespace: kuttl-test-exciting-ghoul === CONT kuttl/harness/es-simple-prod logger.go:42: 07:05:43 | es-simple-prod | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:05:43 | es-simple-prod | Creating namespace: kuttl-test-infinite-moose logger.go:42: 07:05:44 | es-simple-prod | es-simple-prod events from ns kuttl-test-infinite-moose: logger.go:42: 07:05:44 | es-simple-prod | Deleting namespace: kuttl-test-infinite-moose === CONT kuttl/harness/es-rollover-autoprov logger.go:42: 07:05:49 | es-rollover-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:05:49 | es-rollover-autoprov | Creating namespace: kuttl-test-funky-cub logger.go:42: 07:05:50 | es-rollover-autoprov/1-install | starting test step 1-install logger.go:42: 07:05:50 | es-rollover-autoprov/1-install | Jaeger:kuttl-test-funky-cub/my-jaeger created logger.go:42: 07:06:24 | es-rollover-autoprov/1-install | test step completed 1-install logger.go:42: 07:06:24 | es-rollover-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 07:06:24 | es-rollover-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:06:26 | es-rollover-autoprov/2-report-spans | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:06:34 | es-rollover-autoprov/2-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 07:06:34 | es-rollover-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 07:06:34 | es-rollover-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 07:07:00 | es-rollover-autoprov/2-report-spans | test step completed 2-report-spans logger.go:42: 07:07:00 | es-rollover-autoprov/3-check-indices | starting test step 3-check-indices logger.go:42: 07:07:01 | es-rollover-autoprov/3-check-indices | Job:kuttl-test-funky-cub/00-check-indices created logger.go:42: 07:07:06 | es-rollover-autoprov/3-check-indices | test step completed 3-check-indices logger.go:42: 07:07:06 | es-rollover-autoprov/4-check-indices | starting test step 4-check-indices logger.go:42: 07:07:06 | es-rollover-autoprov/4-check-indices | Job:kuttl-test-funky-cub/01-check-indices created logger.go:42: 07:07:12 | es-rollover-autoprov/4-check-indices | test step completed 4-check-indices logger.go:42: 07:07:12 | es-rollover-autoprov/5-install | starting test step 5-install logger.go:42: 07:07:12 | es-rollover-autoprov/5-install | Jaeger:kuttl-test-funky-cub/my-jaeger updated logger.go:42: 07:07:25 | es-rollover-autoprov/5-install | test step completed 5-install logger.go:42: 07:07:25 | es-rollover-autoprov/6-report-spans | starting test step 6-report-spans logger.go:42: 07:07:25 | es-rollover-autoprov/6-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:07:34 | es-rollover-autoprov/6-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=02 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-02-job.yaml] logger.go:42: 07:07:35 | es-rollover-autoprov/6-report-spans | running command: [sh -c kubectl apply -f report-span-02-job.yaml -n $NAMESPACE] logger.go:42: 07:07:35 | es-rollover-autoprov/6-report-spans | job.batch/02-report-span created logger.go:42: 07:08:00 | es-rollover-autoprov/6-report-spans | test step completed 6-report-spans logger.go:42: 07:08:00 | es-rollover-autoprov/7-check-indices | starting test step 7-check-indices logger.go:42: 07:08:00 | es-rollover-autoprov/7-check-indices | Job:kuttl-test-funky-cub/02-check-indices created logger.go:42: 07:08:07 | es-rollover-autoprov/7-check-indices | test step completed 7-check-indices logger.go:42: 07:08:07 | es-rollover-autoprov/8-check-indices | starting test step 8-check-indices logger.go:42: 07:08:07 | es-rollover-autoprov/8-check-indices | Job:kuttl-test-funky-cub/03-check-indices created logger.go:42: 07:08:12 | es-rollover-autoprov/8-check-indices | test step completed 8-check-indices logger.go:42: 07:08:12 | es-rollover-autoprov/9-check-indices | starting test step 9-check-indices logger.go:42: 07:08:13 | es-rollover-autoprov/9-check-indices | Job:kuttl-test-funky-cub/04-check-indices created logger.go:42: 07:08:18 | es-rollover-autoprov/9-check-indices | test step completed 9-check-indices logger.go:42: 07:08:18 | es-rollover-autoprov/10-report-spans | starting test step 10-report-spans logger.go:42: 07:08:18 | es-rollover-autoprov/10-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:08:27 | es-rollover-autoprov/10-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=03 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-03-job.yaml] logger.go:42: 07:08:27 | es-rollover-autoprov/10-report-spans | running command: [sh -c kubectl apply -f report-span-03-job.yaml -n $NAMESPACE] logger.go:42: 07:08:28 | es-rollover-autoprov/10-report-spans | job.batch/03-report-span created logger.go:42: 07:08:54 | es-rollover-autoprov/10-report-spans | test step completed 10-report-spans logger.go:42: 07:08:54 | es-rollover-autoprov/11-check-indices | starting test step 11-check-indices logger.go:42: 07:08:54 | es-rollover-autoprov/11-check-indices | running command: [sh -c go run ../../../../cmd-utils/wait-cronjob/main.go --cronjob my-jaeger-es-rollover --namespace $NAMESPACE] logger.go:42: 07:09:05 | es-rollover-autoprov/11-check-indices | time="2024-09-26T07:09:05Z" level=debug msg="Checking if the my-jaeger-es-rollover CronJob exists" logger.go:42: 07:09:05 | es-rollover-autoprov/11-check-indices | time="2024-09-26T07:09:05Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 07:09:05 | es-rollover-autoprov/11-check-indices | time="2024-09-26T07:09:05Z" level=info msg="Cronjob my-jaeger-es-rollover found successfully" logger.go:42: 07:09:05 | es-rollover-autoprov/11-check-indices | time="2024-09-26T07:09:05Z" level=debug msg="Waiting for the next scheduled job from my-jaeger-es-rollover cronjob" logger.go:42: 07:09:05 | es-rollover-autoprov/11-check-indices | time="2024-09-26T07:09:05Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:09:15 | es-rollover-autoprov/11-check-indices | time="2024-09-26T07:09:15Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:09:25 | es-rollover-autoprov/11-check-indices | time="2024-09-26T07:09:25Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:09:35 | es-rollover-autoprov/11-check-indices | time="2024-09-26T07:09:35Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:09:45 | es-rollover-autoprov/11-check-indices | time="2024-09-26T07:09:45Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:09:55 | es-rollover-autoprov/11-check-indices | time="2024-09-26T07:09:55Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:10:05 | es-rollover-autoprov/11-check-indices | time="2024-09-26T07:10:05Z" level=info msg="Job of owner my-jaeger-es-rollover succeeded after my-jaeger-es-rollover 1m0.170225397s" logger.go:42: 07:10:05 | es-rollover-autoprov/11-check-indices | Job:kuttl-test-funky-cub/05-check-indices created logger.go:42: 07:10:11 | es-rollover-autoprov/11-check-indices | test step completed 11-check-indices logger.go:42: 07:10:11 | es-rollover-autoprov/12-check-indices | starting test step 12-check-indices logger.go:42: 07:10:11 | es-rollover-autoprov/12-check-indices | Job:kuttl-test-funky-cub/06-check-indices created logger.go:42: 07:10:17 | es-rollover-autoprov/12-check-indices | test step completed 12-check-indices logger.go:42: 07:10:17 | es-rollover-autoprov | es-rollover-autoprov events from ns kuttl-test-funky-cub: logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:05:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfunkycubmyjaeger-1-7c655b4f9c-7z47g Binding Scheduled Successfully assigned kuttl-test-funky-cub/elasticsearch-cdm-kuttltestfunkycubmyjaeger-1-7c655b4f9c-7z47g to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:05:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfunkycubmyjaeger-1-7c655b4f9c-7z47g AddedInterface Add eth0 [10.128.0.39/23] from ovn-kubernetes logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:05:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfunkycubmyjaeger-1-7c655b4f9c-7z47g.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:bc97b8e13087050dfb1a75b02c5b780cbb8fb12a849a655169d072ba8bbf42b4" already present on machine kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:05:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfunkycubmyjaeger-1-7c655b4f9c-7z47g.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:05:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfunkycubmyjaeger-1-7c655b4f9c-7z47g.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:05:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfunkycubmyjaeger-1-7c655b4f9c-7z47g.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:1d1d7d47b616995d18692f3a2d6232d7e1a0f41bc1503d87100f31939e080a78" already present on machine kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:05:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfunkycubmyjaeger-1-7c655b4f9c-7z47g.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:05:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfunkycubmyjaeger-1-7c655b4f9c-7z47g.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:05:54 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestfunkycubmyjaeger-1-7c655b4f9c SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestfunkycubmyjaeger-1-7c655b4f9c-7z47g replicaset-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:05:54 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestfunkycubmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestfunkycubmyjaeger-1-7c655b4f9c to 1 deployment-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:04 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestfunkycubmyjaeger-1-7c655b4f9c-7z47g.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:09 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestfunkycubmyjaeger-1-7c655b4f9c-7z47g.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:21 +0000 UTC Normal Pod my-jaeger-collector-7c7d48b746-57clv Binding Scheduled Successfully assigned kuttl-test-funky-cub/my-jaeger-collector-7c7d48b746-57clv to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:21 +0000 UTC Warning Pod my-jaeger-collector-7c7d48b746-57clv FailedMount MountVolume.SetUp failed for volume "my-jaeger-collector-tls-config-volume" : secret "my-jaeger-collector-headless-tls" not found kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:21 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7c7d48b746 SuccessfulCreate Created pod: my-jaeger-collector-7c7d48b746-57clv replicaset-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:21 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-7c7d48b746 to 1 deployment-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:21 +0000 UTC Normal Pod my-jaeger-query-5f4775869b-67zvc Binding Scheduled Successfully assigned kuttl-test-funky-cub/my-jaeger-query-5f4775869b-67zvc to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:21 +0000 UTC Normal Pod my-jaeger-query-5f4775869b-67zvc AddedInterface Add eth0 [10.130.0.40/23] from ovn-kubernetes logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:21 +0000 UTC Normal Pod my-jaeger-query-5f4775869b-67zvc.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:21 +0000 UTC Normal Pod my-jaeger-query-5f4775869b-67zvc.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:21 +0000 UTC Normal Pod my-jaeger-query-5f4775869b-67zvc.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:21 +0000 UTC Normal Pod my-jaeger-query-5f4775869b-67zvc.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:21 +0000 UTC Normal Pod my-jaeger-query-5f4775869b-67zvc.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:21 +0000 UTC Normal Pod my-jaeger-query-5f4775869b-67zvc.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:21 +0000 UTC Normal Pod my-jaeger-query-5f4775869b-67zvc.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:21 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-5f4775869b SuccessfulCreate Created pod: my-jaeger-query-5f4775869b-67zvc replicaset-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:21 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-5f4775869b to 1 deployment-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:22 +0000 UTC Normal Pod my-jaeger-collector-7c7d48b746-57clv AddedInterface Add eth0 [10.129.0.41/23] from ovn-kubernetes logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:22 +0000 UTC Normal Pod my-jaeger-collector-7c7d48b746-57clv.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:22 +0000 UTC Normal Pod my-jaeger-collector-7c7d48b746-57clv.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:22 +0000 UTC Normal Pod my-jaeger-collector-7c7d48b746-57clv.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:22 +0000 UTC Normal Pod my-jaeger-query-5f4775869b-67zvc.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:22 +0000 UTC Normal Pod my-jaeger-query-5f4775869b-67zvc.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:27 +0000 UTC Normal Pod my-jaeger-query-5f4775869b-67zvc.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:27 +0000 UTC Normal Pod my-jaeger-query-5f4775869b-67zvc.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:27 +0000 UTC Normal Pod my-jaeger-query-5f4775869b-67zvc.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:27 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-5f4775869b SuccessfulDelete Deleted pod: my-jaeger-query-5f4775869b-67zvc replicaset-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:27 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-5f4775869b to 0 from 1 deployment-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:28 +0000 UTC Normal Pod my-jaeger-query-5cbb6bd7b6-hbw5j Binding Scheduled Successfully assigned kuttl-test-funky-cub/my-jaeger-query-5cbb6bd7b6-hbw5j to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:28 +0000 UTC Normal Pod my-jaeger-query-5cbb6bd7b6-hbw5j AddedInterface Add eth0 [10.130.0.41/23] from ovn-kubernetes logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:28 +0000 UTC Normal Pod my-jaeger-query-5cbb6bd7b6-hbw5j.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:28 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-5cbb6bd7b6 SuccessfulCreate Created pod: my-jaeger-query-5cbb6bd7b6-hbw5j replicaset-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:28 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-5cbb6bd7b6 to 1 deployment-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:29 +0000 UTC Normal Pod my-jaeger-query-5cbb6bd7b6-hbw5j.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:29 +0000 UTC Normal Pod my-jaeger-query-5cbb6bd7b6-hbw5j.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:29 +0000 UTC Normal Pod my-jaeger-query-5cbb6bd7b6-hbw5j.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:29 +0000 UTC Normal Pod my-jaeger-query-5cbb6bd7b6-hbw5j.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:29 +0000 UTC Normal Pod my-jaeger-query-5cbb6bd7b6-hbw5j.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:29 +0000 UTC Normal Pod my-jaeger-query-5cbb6bd7b6-hbw5j.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:29 +0000 UTC Normal Pod my-jaeger-query-5cbb6bd7b6-hbw5j.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:29 +0000 UTC Normal Pod my-jaeger-query-5cbb6bd7b6-hbw5j.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:34 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-p5hml job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:35 +0000 UTC Normal Pod 00-report-span-p5hml Binding Scheduled Successfully assigned kuttl-test-funky-cub/00-report-span-p5hml to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:35 +0000 UTC Normal Pod 00-report-span-p5hml AddedInterface Add eth0 [10.129.0.42/23] from ovn-kubernetes logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:35 +0000 UTC Normal Pod 00-report-span-p5hml.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:36 +0000 UTC Normal Pod 00-report-span-p5hml.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 929ms (929ms including waiting) kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:36 +0000 UTC Normal Pod 00-report-span-p5hml.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:36 +0000 UTC Normal Pod 00-report-span-p5hml.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:36 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:36 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:36 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:06:59 +0000 UTC Normal Job.batch 00-report-span Completed Job completed job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:01 +0000 UTC Normal Pod 00-check-indices-55jjb Binding Scheduled Successfully assigned kuttl-test-funky-cub/00-check-indices-55jjb to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:01 +0000 UTC Normal Pod 00-check-indices-55jjb AddedInterface Add eth0 [10.129.0.43/23] from ovn-kubernetes logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:01 +0000 UTC Normal Pod 00-check-indices-55jjb.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:01 +0000 UTC Normal Job.batch 00-check-indices SuccessfulCreate Created pod: 00-check-indices-55jjb job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:02 +0000 UTC Normal Pod 00-check-indices-55jjb.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 945ms (945ms including waiting) kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:02 +0000 UTC Normal Pod 00-check-indices-55jjb.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:02 +0000 UTC Normal Pod 00-check-indices-55jjb.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:05 +0000 UTC Normal Job.batch 00-check-indices Completed Job completed job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:06 +0000 UTC Normal Pod 01-check-indices-qzf6p Binding Scheduled Successfully assigned kuttl-test-funky-cub/01-check-indices-qzf6p to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:06 +0000 UTC Normal Job.batch 01-check-indices SuccessfulCreate Created pod: 01-check-indices-qzf6p job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:06 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-7c7d48b746-57clv horizontal-pod-autoscaler logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:07 +0000 UTC Normal Pod 01-check-indices-qzf6p AddedInterface Add eth0 [10.129.0.44/23] from ovn-kubernetes logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:07 +0000 UTC Normal Pod 01-check-indices-qzf6p.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:08 +0000 UTC Normal Pod 01-check-indices-qzf6p.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 839ms (839ms including waiting) kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:08 +0000 UTC Normal Pod 01-check-indices-qzf6p.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:08 +0000 UTC Normal Pod 01-check-indices-qzf6p.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:11 +0000 UTC Normal Job.batch 01-check-indices Completed Job completed job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:12 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-tdg7m Binding Scheduled Successfully assigned kuttl-test-funky-cub/my-jaeger-es-rollover-create-mapping-tdg7m to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:12 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping SuccessfulCreate Created pod: my-jaeger-es-rollover-create-mapping-tdg7m job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:13 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-tdg7m AddedInterface Add eth0 [10.129.0.45/23] from ovn-kubernetes logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:13 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-tdg7m.spec.containers{my-jaeger-es-rollover-create-mapping} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:b3897239050ab12ae12c28bb51d955438334aad957dceb83a53cf2e8a466b69b" kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:20 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-tdg7m.spec.containers{my-jaeger-es-rollover-create-mapping} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:b3897239050ab12ae12c28bb51d955438334aad957dceb83a53cf2e8a466b69b" in 7.11s (7.11s including waiting) kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:20 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-tdg7m.spec.containers{my-jaeger-es-rollover-create-mapping} Created Created container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:20 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-tdg7m.spec.containers{my-jaeger-es-rollover-create-mapping} Started Started container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:21 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:21 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:24 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping Completed Job completed job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:25 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-7c7d48b746 to 0 from 1 deployment-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:26 +0000 UTC Normal Pod my-jaeger-collector-5c9c58c57-54hw4 Binding Scheduled Successfully assigned kuttl-test-funky-cub/my-jaeger-collector-5c9c58c57-54hw4 to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:26 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-5c9c58c57 SuccessfulCreate Created pod: my-jaeger-collector-5c9c58c57-54hw4 replicaset-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:26 +0000 UTC Normal Pod my-jaeger-collector-7c7d48b746-57clv.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:26 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7c7d48b746 SuccessfulDelete Deleted pod: my-jaeger-collector-7c7d48b746-57clv replicaset-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:26 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-5c9c58c57 to 1 deployment-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:26 +0000 UTC Normal Pod my-jaeger-query-5cbb6bd7b6-hbw5j.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:26 +0000 UTC Normal Pod my-jaeger-query-5cbb6bd7b6-hbw5j.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:26 +0000 UTC Normal Pod my-jaeger-query-5cbb6bd7b6-hbw5j.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:26 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-5cbb6bd7b6 SuccessfulDelete Deleted pod: my-jaeger-query-5cbb6bd7b6-hbw5j replicaset-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:26 +0000 UTC Normal Pod my-jaeger-query-848b5664f5-jtd4z Binding Scheduled Successfully assigned kuttl-test-funky-cub/my-jaeger-query-848b5664f5-jtd4z to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:26 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-848b5664f5 SuccessfulCreate Created pod: my-jaeger-query-848b5664f5-jtd4z replicaset-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:26 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-5cbb6bd7b6 to 0 from 1 deployment-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:26 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-848b5664f5 to 1 deployment-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:27 +0000 UTC Normal Pod my-jaeger-collector-5c9c58c57-54hw4 AddedInterface Add eth0 [10.129.0.46/23] from ovn-kubernetes logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:27 +0000 UTC Normal Pod my-jaeger-collector-5c9c58c57-54hw4.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:27 +0000 UTC Normal Pod my-jaeger-collector-5c9c58c57-54hw4.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:27 +0000 UTC Normal Pod my-jaeger-collector-5c9c58c57-54hw4.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:27 +0000 UTC Normal Pod my-jaeger-query-848b5664f5-jtd4z AddedInterface Add eth0 [10.130.0.42/23] from ovn-kubernetes logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:27 +0000 UTC Normal Pod my-jaeger-query-848b5664f5-jtd4z.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:27 +0000 UTC Normal Pod my-jaeger-query-848b5664f5-jtd4z.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:27 +0000 UTC Normal Pod my-jaeger-query-848b5664f5-jtd4z.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:27 +0000 UTC Normal Pod my-jaeger-query-848b5664f5-jtd4z.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:27 +0000 UTC Normal Pod my-jaeger-query-848b5664f5-jtd4z.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:27 +0000 UTC Normal Pod my-jaeger-query-848b5664f5-jtd4z.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:27 +0000 UTC Normal Pod my-jaeger-query-848b5664f5-jtd4z.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:27 +0000 UTC Normal Pod my-jaeger-query-848b5664f5-jtd4z.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:27 +0000 UTC Normal Pod my-jaeger-query-848b5664f5-jtd4z.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:35 +0000 UTC Normal Pod 02-report-span-kx5xz Binding Scheduled Successfully assigned kuttl-test-funky-cub/02-report-span-kx5xz to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:35 +0000 UTC Normal Job.batch 02-report-span SuccessfulCreate Created pod: 02-report-span-kx5xz job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:36 +0000 UTC Normal Pod 02-report-span-kx5xz AddedInterface Add eth0 [10.129.0.47/23] from ovn-kubernetes logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:36 +0000 UTC Normal Pod 02-report-span-kx5xz.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:36 +0000 UTC Normal Pod 02-report-span-kx5xz.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 826ms (826ms including waiting) kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:36 +0000 UTC Normal Pod 02-report-span-kx5xz.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:36 +0000 UTC Normal Pod 02-report-span-kx5xz.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:07:59 +0000 UTC Normal Job.batch 02-report-span Completed Job completed job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:00 +0000 UTC Normal Pod 02-check-indices-hpdjn Binding Scheduled Successfully assigned kuttl-test-funky-cub/02-check-indices-hpdjn to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:00 +0000 UTC Normal Pod 02-check-indices-hpdjn AddedInterface Add eth0 [10.129.0.50/23] from ovn-kubernetes logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:00 +0000 UTC Normal Job.batch 02-check-indices SuccessfulCreate Created pod: 02-check-indices-hpdjn job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28788908-drqr2 Binding Scheduled Successfully assigned kuttl-test-funky-cub/my-jaeger-es-lookback-28788908-drqr2 to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28788908-drqr2 AddedInterface Add eth0 [10.129.0.48/23] from ovn-kubernetes logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28788908-drqr2.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:b3897239050ab12ae12c28bb51d955438334aad957dceb83a53cf2e8a466b69b" already present on machine kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28788908-drqr2.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28788908-drqr2.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28788908 SuccessfulCreate Created pod: my-jaeger-es-lookback-28788908-drqr2 job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28788908 cronjob-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28788908-p52v6 Binding Scheduled Successfully assigned kuttl-test-funky-cub/my-jaeger-es-rollover-28788908-p52v6 to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28788908-p52v6 AddedInterface Add eth0 [10.129.0.49/23] from ovn-kubernetes logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28788908-p52v6.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:b3897239050ab12ae12c28bb51d955438334aad957dceb83a53cf2e8a466b69b" already present on machine kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28788908-p52v6.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28788908-p52v6.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28788908 SuccessfulCreate Created pod: my-jaeger-es-rollover-28788908-p52v6 job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28788908 cronjob-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:01 +0000 UTC Normal Pod 02-check-indices-hpdjn.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:01 +0000 UTC Normal Pod 02-check-indices-hpdjn.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 867ms (867ms including waiting) kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:02 +0000 UTC Normal Pod 02-check-indices-hpdjn.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:02 +0000 UTC Normal Pod 02-check-indices-hpdjn.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:04 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28788908 Completed Job completed job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:04 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28788908, status: Complete cronjob-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:04 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28788908 Completed Job completed job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:04 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28788908, status: Complete cronjob-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:05 +0000 UTC Normal Job.batch 02-check-indices Completed Job completed job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:07 +0000 UTC Normal Pod 03-check-indices-bnbrg Binding Scheduled Successfully assigned kuttl-test-funky-cub/03-check-indices-bnbrg to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:07 +0000 UTC Normal Pod 03-check-indices-bnbrg AddedInterface Add eth0 [10.129.0.51/23] from ovn-kubernetes logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:07 +0000 UTC Normal Pod 03-check-indices-bnbrg.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:07 +0000 UTC Normal Job.batch 03-check-indices SuccessfulCreate Created pod: 03-check-indices-bnbrg job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:08 +0000 UTC Normal Pod 03-check-indices-bnbrg.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 828ms (828ms including waiting) kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:08 +0000 UTC Normal Pod 03-check-indices-bnbrg.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:08 +0000 UTC Normal Pod 03-check-indices-bnbrg.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:12 +0000 UTC Normal Job.batch 03-check-indices Completed Job completed job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:13 +0000 UTC Normal Pod 04-check-indices-xz6p2 Binding Scheduled Successfully assigned kuttl-test-funky-cub/04-check-indices-xz6p2 to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:13 +0000 UTC Normal Pod 04-check-indices-xz6p2 AddedInterface Add eth0 [10.129.0.52/23] from ovn-kubernetes logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:13 +0000 UTC Normal Pod 04-check-indices-xz6p2.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:13 +0000 UTC Normal Job.batch 04-check-indices SuccessfulCreate Created pod: 04-check-indices-xz6p2 job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:14 +0000 UTC Normal Pod 04-check-indices-xz6p2.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 853ms (853ms including waiting) kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:14 +0000 UTC Normal Pod 04-check-indices-xz6p2.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:14 +0000 UTC Normal Pod 04-check-indices-xz6p2.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:18 +0000 UTC Normal Job.batch 04-check-indices Completed Job completed job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:28 +0000 UTC Normal Pod 03-report-span-54g2f Binding Scheduled Successfully assigned kuttl-test-funky-cub/03-report-span-54g2f to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:28 +0000 UTC Normal Pod 03-report-span-54g2f AddedInterface Add eth0 [10.129.0.53/23] from ovn-kubernetes logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:28 +0000 UTC Normal Pod 03-report-span-54g2f.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:28 +0000 UTC Normal Job.batch 03-report-span SuccessfulCreate Created pod: 03-report-span-54g2f job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:29 +0000 UTC Normal Pod 03-report-span-54g2f.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 806ms (806ms including waiting) kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:29 +0000 UTC Normal Pod 03-report-span-54g2f.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:29 +0000 UTC Normal Pod 03-report-span-54g2f.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:08:53 +0000 UTC Normal Job.batch 03-report-span Completed Job completed job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:09:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28788909-4s5tw Binding Scheduled Successfully assigned kuttl-test-funky-cub/my-jaeger-es-lookback-28788909-4s5tw to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:09:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28788909-4s5tw AddedInterface Add eth0 [10.129.0.55/23] from ovn-kubernetes logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:09:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28788909-4s5tw.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:b3897239050ab12ae12c28bb51d955438334aad957dceb83a53cf2e8a466b69b" already present on machine kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:09:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28788909-4s5tw.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:09:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28788909-4s5tw.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:09:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28788909 SuccessfulCreate Created pod: my-jaeger-es-lookback-28788909-4s5tw job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:09:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28788909 cronjob-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:09:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28788909-x7hmv Binding Scheduled Successfully assigned kuttl-test-funky-cub/my-jaeger-es-rollover-28788909-x7hmv to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:09:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28788909-x7hmv AddedInterface Add eth0 [10.129.0.54/23] from ovn-kubernetes logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:09:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28788909-x7hmv.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:b3897239050ab12ae12c28bb51d955438334aad957dceb83a53cf2e8a466b69b" already present on machine kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:09:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28788909-x7hmv.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:09:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28788909-x7hmv.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:09:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28788909 SuccessfulCreate Created pod: my-jaeger-es-rollover-28788909-x7hmv job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:09:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28788909 cronjob-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:09:04 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28788909 Completed Job completed job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:09:04 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28788909, status: Complete cronjob-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:09:04 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28788909 Completed Job completed job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:09:04 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28788909, status: Complete cronjob-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28788910-s264f Binding Scheduled Successfully assigned kuttl-test-funky-cub/my-jaeger-es-lookback-28788910-s264f to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28788910-s264f AddedInterface Add eth0 [10.129.0.56/23] from ovn-kubernetes logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28788910-s264f.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:b3897239050ab12ae12c28bb51d955438334aad957dceb83a53cf2e8a466b69b" already present on machine kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28788910-s264f.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28788910-s264f.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28788910 SuccessfulCreate Created pod: my-jaeger-es-lookback-28788910-s264f job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28788910 cronjob-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28788910-4cs8w Binding Scheduled Successfully assigned kuttl-test-funky-cub/my-jaeger-es-rollover-28788910-4cs8w to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28788910-4cs8w AddedInterface Add eth0 [10.129.0.57/23] from ovn-kubernetes logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28788910-4cs8w.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:b3897239050ab12ae12c28bb51d955438334aad957dceb83a53cf2e8a466b69b" already present on machine kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28788910-4cs8w.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28788910-4cs8w.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28788910 SuccessfulCreate Created pod: my-jaeger-es-rollover-28788910-4cs8w job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28788910 cronjob-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:04 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28788910 Completed Job completed job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:04 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28788910, status: Complete cronjob-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:04 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28788910 Completed Job completed job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:04 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28788910, status: Complete cronjob-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:05 +0000 UTC Normal Pod 05-check-indices-8cj7q Binding Scheduled Successfully assigned kuttl-test-funky-cub/05-check-indices-8cj7q to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:05 +0000 UTC Normal Job.batch 05-check-indices SuccessfulCreate Created pod: 05-check-indices-8cj7q job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:06 +0000 UTC Normal Pod 05-check-indices-8cj7q AddedInterface Add eth0 [10.129.0.58/23] from ovn-kubernetes logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:06 +0000 UTC Normal Pod 05-check-indices-8cj7q.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:07 +0000 UTC Normal Pod 05-check-indices-8cj7q.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 823ms (823ms including waiting) kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:07 +0000 UTC Normal Pod 05-check-indices-8cj7q.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:07 +0000 UTC Normal Pod 05-check-indices-8cj7q.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:10 +0000 UTC Normal Job.batch 05-check-indices Completed Job completed job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:11 +0000 UTC Normal Pod 06-check-indices-fcc7p Binding Scheduled Successfully assigned kuttl-test-funky-cub/06-check-indices-fcc7p to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:11 +0000 UTC Normal Job.batch 06-check-indices SuccessfulCreate Created pod: 06-check-indices-fcc7p job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:12 +0000 UTC Normal Pod 06-check-indices-fcc7p AddedInterface Add eth0 [10.129.0.59/23] from ovn-kubernetes logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:12 +0000 UTC Normal Pod 06-check-indices-fcc7p.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:13 +0000 UTC Normal Pod 06-check-indices-fcc7p.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 805ms (805ms including waiting) kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:13 +0000 UTC Normal Pod 06-check-indices-fcc7p.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:13 +0000 UTC Normal Pod 06-check-indices-fcc7p.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:10:17 | es-rollover-autoprov | 2024-09-26 07:10:16 +0000 UTC Normal Job.batch 06-check-indices Completed Job completed job-controller logger.go:42: 07:10:17 | es-rollover-autoprov | Deleting namespace: kuttl-test-funky-cub === CONT kuttl/harness/es-increasing-replicas logger.go:42: 07:10:25 | es-increasing-replicas | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:10:25 | es-increasing-replicas | Ignoring check-es-nodes.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:10:25 | es-increasing-replicas | Ignoring openshift-check-es-nodes.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:10:25 | es-increasing-replicas | Creating namespace: kuttl-test-well-unicorn logger.go:42: 07:10:25 | es-increasing-replicas/1-install | starting test step 1-install logger.go:42: 07:10:25 | es-increasing-replicas/1-install | Jaeger:kuttl-test-well-unicorn/simple-prod created logger.go:42: 07:11:02 | es-increasing-replicas/1-install | test step completed 1-install logger.go:42: 07:11:02 | es-increasing-replicas/2-install | starting test step 2-install logger.go:42: 07:11:02 | es-increasing-replicas/2-install | Jaeger:kuttl-test-well-unicorn/simple-prod updated logger.go:42: 07:11:12 | es-increasing-replicas/2-install | test step completed 2-install logger.go:42: 07:11:12 | es-increasing-replicas/3-smoke-test | starting test step 3-smoke-test logger.go:42: 07:11:12 | es-increasing-replicas/3-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:11:14 | es-increasing-replicas/3-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:11:29 | es-increasing-replicas/3-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:11:30 | es-increasing-replicas/3-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:11:30 | es-increasing-replicas/3-smoke-test | job.batch/report-span created logger.go:42: 07:11:30 | es-increasing-replicas/3-smoke-test | job.batch/check-span created logger.go:42: 07:11:43 | es-increasing-replicas/3-smoke-test | test step completed 3-smoke-test logger.go:42: 07:11:43 | es-increasing-replicas/4-install | starting test step 4-install logger.go:42: 07:11:44 | es-increasing-replicas/4-install | Jaeger:kuttl-test-well-unicorn/simple-prod updated logger.go:42: 07:11:44 | es-increasing-replicas/4-install | test step completed 4-install logger.go:42: 07:11:44 | es-increasing-replicas/5-check-es-nodes | starting test step 5-check-es-nodes logger.go:42: 07:11:44 | es-increasing-replicas/5-check-es-nodes | running command: [sh -c ./check-es-nodes.sh $NAMESPACE] logger.go:42: 07:11:44 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 07:11:44 | es-increasing-replicas/5-check-es-nodes | false logger.go:42: 07:11:44 | es-increasing-replicas/5-check-es-nodes | Error: no matches found logger.go:42: 07:11:49 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 07:11:49 | es-increasing-replicas/5-check-es-nodes | true logger.go:42: 07:11:49 | es-increasing-replicas/5-check-es-nodes | test step completed 5-check-es-nodes logger.go:42: 07:11:50 | es-increasing-replicas | es-increasing-replicas events from ns kuttl-test-well-unicorn: logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:31 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestwellunicornsimpleprod-1-6f9fdd9769 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestwellunicornsimpleprod-1-6f9fdd99tmw6 replicaset-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:31 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwellunicornsimpleprod-1-6f9fdd99tmw6 Binding Scheduled Successfully assigned kuttl-test-well-unicorn/elasticsearch-cdm-kuttltestwellunicornsimpleprod-1-6f9fdd99tmw6 to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:31 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwellunicornsimpleprod-1-6f9fdd99tmw6 AddedInterface Add eth0 [10.128.0.40/23] from ovn-kubernetes logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:31 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwellunicornsimpleprod-1-6f9fdd99tmw6.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:bc97b8e13087050dfb1a75b02c5b780cbb8fb12a849a655169d072ba8bbf42b4" already present on machine kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:31 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwellunicornsimpleprod-1-6f9fdd99tmw6.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:31 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwellunicornsimpleprod-1-6f9fdd99tmw6.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:31 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwellunicornsimpleprod-1-6f9fdd99tmw6.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:1d1d7d47b616995d18692f3a2d6232d7e1a0f41bc1503d87100f31939e080a78" already present on machine kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:31 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwellunicornsimpleprod-1-6f9fdd99tmw6.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:31 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwellunicornsimpleprod-1-6f9fdd99tmw6.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:31 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestwellunicornsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestwellunicornsimpleprod-1-6f9fdd9769 to 1 deployment-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:41 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestwellunicornsimpleprod-1-6f9fdd99tmw6.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:46 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestwellunicornsimpleprod-1-6f9fdd99tmw6.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:58 +0000 UTC Normal Pod simple-prod-collector-bbddf5775-jznkg Binding Scheduled Successfully assigned kuttl-test-well-unicorn/simple-prod-collector-bbddf5775-jznkg to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:58 +0000 UTC Normal Pod simple-prod-collector-bbddf5775-jznkg AddedInterface Add eth0 [10.129.0.60/23] from ovn-kubernetes logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:58 +0000 UTC Normal Pod simple-prod-collector-bbddf5775-jznkg.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:58 +0000 UTC Normal Pod simple-prod-collector-bbddf5775-jznkg.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:58 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-bbddf5775 SuccessfulCreate Created pod: simple-prod-collector-bbddf5775-jznkg replicaset-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:58 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-bbddf5775 to 1 deployment-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:58 +0000 UTC Normal Pod simple-prod-query-85b859f9b7-xlwns Binding Scheduled Successfully assigned kuttl-test-well-unicorn/simple-prod-query-85b859f9b7-xlwns to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:58 +0000 UTC Warning Pod simple-prod-query-85b859f9b7-xlwns FailedMount MountVolume.SetUp failed for volume "simple-prod-ui-oauth-proxy-tls" : secret "simple-prod-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:58 +0000 UTC Normal ReplicaSet.apps simple-prod-query-85b859f9b7 SuccessfulCreate Created pod: simple-prod-query-85b859f9b7-xlwns replicaset-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:58 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-85b859f9b7 to 1 deployment-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:59 +0000 UTC Normal Pod simple-prod-collector-bbddf5775-jznkg.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:59 +0000 UTC Normal Pod simple-prod-query-85b859f9b7-xlwns AddedInterface Add eth0 [10.130.0.44/23] from ovn-kubernetes logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:59 +0000 UTC Normal Pod simple-prod-query-85b859f9b7-xlwns.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:59 +0000 UTC Normal Pod simple-prod-query-85b859f9b7-xlwns.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:59 +0000 UTC Normal Pod simple-prod-query-85b859f9b7-xlwns.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:59 +0000 UTC Normal Pod simple-prod-query-85b859f9b7-xlwns.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:59 +0000 UTC Normal Pod simple-prod-query-85b859f9b7-xlwns.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:59 +0000 UTC Normal Pod simple-prod-query-85b859f9b7-xlwns.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:59 +0000 UTC Normal Pod simple-prod-query-85b859f9b7-xlwns.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:59 +0000 UTC Normal Pod simple-prod-query-85b859f9b7-xlwns.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:10:59 +0000 UTC Normal Pod simple-prod-query-85b859f9b7-xlwns.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:03 +0000 UTC Normal Pod simple-prod-collector-bbddf5775-q6hsf Binding Scheduled Successfully assigned kuttl-test-well-unicorn/simple-prod-collector-bbddf5775-q6hsf to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:03 +0000 UTC Normal Pod simple-prod-collector-bbddf5775-q6hsf AddedInterface Add eth0 [10.130.0.45/23] from ovn-kubernetes logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:03 +0000 UTC Normal Pod simple-prod-collector-bbddf5775-q6hsf.spec.containers{jaeger-collector} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:03 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-bbddf5775 SuccessfulCreate Created pod: simple-prod-collector-bbddf5775-q6hsf replicaset-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:03 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-bbddf5775 to 2 from 1 deployment-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:03 +0000 UTC Normal Pod simple-prod-query-85b859f9b7-gsbf8 Binding Scheduled Successfully assigned kuttl-test-well-unicorn/simple-prod-query-85b859f9b7-gsbf8 to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:03 +0000 UTC Normal Pod simple-prod-query-85b859f9b7-gsbf8 AddedInterface Add eth0 [10.128.0.41/23] from ovn-kubernetes logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:03 +0000 UTC Normal ReplicaSet.apps simple-prod-query-85b859f9b7 SuccessfulCreate Created pod: simple-prod-query-85b859f9b7-gsbf8 replicaset-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:03 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-85b859f9b7 to 2 from 1 deployment-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:04 +0000 UTC Normal Pod simple-prod-query-85b859f9b7-gsbf8.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:06 +0000 UTC Normal Pod simple-prod-collector-bbddf5775-q6hsf.spec.containers{jaeger-collector} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" in 2.839s (2.839s including waiting) kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:06 +0000 UTC Normal Pod simple-prod-collector-bbddf5775-q6hsf.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:06 +0000 UTC Normal Pod simple-prod-collector-bbddf5775-q6hsf.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:07 +0000 UTC Normal Pod simple-prod-query-85b859f9b7-gsbf8.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" in 3.784s (3.784s including waiting) kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:07 +0000 UTC Normal Pod simple-prod-query-85b859f9b7-gsbf8.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:07 +0000 UTC Normal Pod simple-prod-query-85b859f9b7-gsbf8.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:07 +0000 UTC Normal Pod simple-prod-query-85b859f9b7-gsbf8.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:07 +0000 UTC Normal Pod simple-prod-query-85b859f9b7-gsbf8.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:08 +0000 UTC Normal Pod simple-prod-query-85b859f9b7-gsbf8.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:08 +0000 UTC Normal Pod simple-prod-query-85b859f9b7-gsbf8.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:10 +0000 UTC Normal Pod simple-prod-query-85b859f9b7-gsbf8.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" in 2.508s (2.508s including waiting) kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:10 +0000 UTC Normal Pod simple-prod-query-85b859f9b7-gsbf8.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:10 +0000 UTC Normal Pod simple-prod-query-85b859f9b7-gsbf8.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:15 +0000 UTC Normal Pod simple-prod-query-85b859f9b7-gsbf8.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:15 +0000 UTC Normal Pod simple-prod-query-85b859f9b7-gsbf8.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:15 +0000 UTC Normal Pod simple-prod-query-85b859f9b7-gsbf8.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:15 +0000 UTC Normal Pod simple-prod-query-85b859f9b7-xlwns.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:15 +0000 UTC Normal Pod simple-prod-query-85b859f9b7-xlwns.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:15 +0000 UTC Normal Pod simple-prod-query-85b859f9b7-xlwns.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:15 +0000 UTC Normal ReplicaSet.apps simple-prod-query-85b859f9b7 SuccessfulDelete Deleted pod: simple-prod-query-85b859f9b7-gsbf8 replicaset-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:15 +0000 UTC Normal ReplicaSet.apps simple-prod-query-85b859f9b7 SuccessfulDelete Deleted pod: simple-prod-query-85b859f9b7-xlwns replicaset-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:15 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-85b859f9b7 to 0 from 2 deployment-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:17 +0000 UTC Normal Pod simple-prod-query-6dcfb76c98-22wj6 Binding Scheduled Successfully assigned kuttl-test-well-unicorn/simple-prod-query-6dcfb76c98-22wj6 to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:17 +0000 UTC Normal Pod simple-prod-query-6dcfb76c98-22wj6 AddedInterface Add eth0 [10.129.0.61/23] from ovn-kubernetes logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:17 +0000 UTC Normal Pod simple-prod-query-6dcfb76c98-wfltb Binding Scheduled Successfully assigned kuttl-test-well-unicorn/simple-prod-query-6dcfb76c98-wfltb to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:17 +0000 UTC Normal Pod simple-prod-query-6dcfb76c98-wfltb AddedInterface Add eth0 [10.130.0.46/23] from ovn-kubernetes logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:17 +0000 UTC Normal Pod simple-prod-query-6dcfb76c98-wfltb.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:17 +0000 UTC Normal Pod simple-prod-query-6dcfb76c98-wfltb.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:17 +0000 UTC Normal Pod simple-prod-query-6dcfb76c98-wfltb.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:17 +0000 UTC Normal Pod simple-prod-query-6dcfb76c98-wfltb.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:17 +0000 UTC Normal Pod simple-prod-query-6dcfb76c98-wfltb.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:17 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6dcfb76c98 SuccessfulCreate Created pod: simple-prod-query-6dcfb76c98-wfltb replicaset-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:17 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6dcfb76c98 SuccessfulCreate Created pod: simple-prod-query-6dcfb76c98-22wj6 replicaset-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:17 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-6dcfb76c98 to 2 deployment-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:18 +0000 UTC Normal Pod simple-prod-query-6dcfb76c98-22wj6.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:18 +0000 UTC Normal Pod simple-prod-query-6dcfb76c98-wfltb.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:18 +0000 UTC Normal Pod simple-prod-query-6dcfb76c98-wfltb.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:18 +0000 UTC Normal Pod simple-prod-query-6dcfb76c98-wfltb.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:18 +0000 UTC Normal Pod simple-prod-query-6dcfb76c98-wfltb.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:21 +0000 UTC Normal Pod simple-prod-query-6dcfb76c98-22wj6.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" in 3.296s (3.296s including waiting) kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:21 +0000 UTC Normal Pod simple-prod-query-6dcfb76c98-22wj6.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:21 +0000 UTC Normal Pod simple-prod-query-6dcfb76c98-22wj6.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:21 +0000 UTC Normal Pod simple-prod-query-6dcfb76c98-22wj6.spec.containers{oauth-proxy} Pulling Pulling image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:24 +0000 UTC Normal Pod simple-prod-query-6dcfb76c98-22wj6.spec.containers{oauth-proxy} Pulled Successfully pulled image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" in 3.513s (3.514s including waiting) kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:25 +0000 UTC Normal Pod simple-prod-query-6dcfb76c98-22wj6.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:25 +0000 UTC Normal Pod simple-prod-query-6dcfb76c98-22wj6.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:25 +0000 UTC Normal Pod simple-prod-query-6dcfb76c98-22wj6.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:27 +0000 UTC Normal Pod simple-prod-query-6dcfb76c98-22wj6.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" in 2.519s (2.519s including waiting) kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:27 +0000 UTC Normal Pod simple-prod-query-6dcfb76c98-22wj6.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:27 +0000 UTC Normal Pod simple-prod-query-6dcfb76c98-22wj6.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:30 +0000 UTC Normal Pod check-span-rghc9 Binding Scheduled Successfully assigned kuttl-test-well-unicorn/check-span-rghc9 to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:30 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-rghc9 job-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:30 +0000 UTC Normal Pod report-span-74jpt Binding Scheduled Successfully assigned kuttl-test-well-unicorn/report-span-74jpt to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:30 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-74jpt job-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:31 +0000 UTC Normal Pod check-span-rghc9 AddedInterface Add eth0 [10.129.0.62/23] from ovn-kubernetes logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:31 +0000 UTC Normal Pod check-span-rghc9.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:31 +0000 UTC Normal Pod report-span-74jpt AddedInterface Add eth0 [10.130.0.47/23] from ovn-kubernetes logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:31 +0000 UTC Normal Pod report-span-74jpt.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:31 +0000 UTC Normal Pod report-span-74jpt.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 850ms (850ms including waiting) kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:32 +0000 UTC Normal Pod check-span-rghc9.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 838ms (838ms including waiting) kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:32 +0000 UTC Normal Pod check-span-rghc9.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:32 +0000 UTC Normal Pod check-span-rghc9.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:32 +0000 UTC Normal Pod report-span-74jpt.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:32 +0000 UTC Normal Pod report-span-74jpt.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:43 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:44 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestwellunicornsimpleprod-2-dc7ff9d5c SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestwellunicornsimpleprod-2-dc7ff9dzq8k7 replicaset-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:44 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwellunicornsimpleprod-2-dc7ff9dzq8k7 Binding Scheduled Successfully assigned kuttl-test-well-unicorn/elasticsearch-cdm-kuttltestwellunicornsimpleprod-2-dc7ff9dzq8k7 to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:44 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestwellunicornsimpleprod-2 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestwellunicornsimpleprod-2-dc7ff9d5c to 1 deployment-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:44 +0000 UTC Normal Pod simple-prod-collector-bbddf5775-jznkg.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:44 +0000 UTC Normal Pod simple-prod-collector-bbddf5775-q6hsf.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:44 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-bbddf5775 SuccessfulDelete Deleted pod: simple-prod-collector-bbddf5775-q6hsf replicaset-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:44 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-bbddf5775 SuccessfulDelete Deleted pod: simple-prod-collector-bbddf5775-jznkg replicaset-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:44 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled down replica set simple-prod-collector-bbddf5775 to 0 from 2 deployment-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:44 +0000 UTC Normal Pod simple-prod-query-6dcfb76c98-22wj6.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:44 +0000 UTC Normal Pod simple-prod-query-6dcfb76c98-22wj6.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:44 +0000 UTC Normal Pod simple-prod-query-6dcfb76c98-22wj6.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:44 +0000 UTC Normal Pod simple-prod-query-6dcfb76c98-wfltb.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:44 +0000 UTC Normal Pod simple-prod-query-6dcfb76c98-wfltb.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:44 +0000 UTC Normal Pod simple-prod-query-6dcfb76c98-wfltb.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:44 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6dcfb76c98 SuccessfulDelete Deleted pod: simple-prod-query-6dcfb76c98-22wj6 replicaset-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:44 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6dcfb76c98 SuccessfulDelete Deleted pod: simple-prod-query-6dcfb76c98-wfltb replicaset-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:44 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-6dcfb76c98 to 0 from 2 deployment-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwellunicornsimpleprod-2-dc7ff9dzq8k7 AddedInterface Add eth0 [10.129.0.63/23] from ovn-kubernetes logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:45 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestwellunicornsimpleprod-2-dc7ff9dzq8k7.spec.containers{elasticsearch} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:bc97b8e13087050dfb1a75b02c5b780cbb8fb12a849a655169d072ba8bbf42b4" kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:45 +0000 UTC Normal Pod simple-prod-collector-5bf6cc7bcb-9s9bg Binding Scheduled Successfully assigned kuttl-test-well-unicorn/simple-prod-collector-5bf6cc7bcb-9s9bg to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:45 +0000 UTC Normal Pod simple-prod-collector-5bf6cc7bcb-wkkzn Binding Scheduled Successfully assigned kuttl-test-well-unicorn/simple-prod-collector-5bf6cc7bcb-wkkzn to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:45 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-5bf6cc7bcb SuccessfulCreate Created pod: simple-prod-collector-5bf6cc7bcb-wkkzn replicaset-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:45 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-5bf6cc7bcb SuccessfulCreate Created pod: simple-prod-collector-5bf6cc7bcb-9s9bg replicaset-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:45 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-5bf6cc7bcb to 2 deployment-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:45 +0000 UTC Normal Pod simple-prod-query-78fbc95b7-fxx66 Binding Scheduled Successfully assigned kuttl-test-well-unicorn/simple-prod-query-78fbc95b7-fxx66 to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:45 +0000 UTC Normal Pod simple-prod-query-78fbc95b7-jchz5 Binding Scheduled Successfully assigned kuttl-test-well-unicorn/simple-prod-query-78fbc95b7-jchz5 to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:45 +0000 UTC Normal ReplicaSet.apps simple-prod-query-78fbc95b7 SuccessfulCreate Created pod: simple-prod-query-78fbc95b7-jchz5 replicaset-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:45 +0000 UTC Normal ReplicaSet.apps simple-prod-query-78fbc95b7 SuccessfulCreate Created pod: simple-prod-query-78fbc95b7-fxx66 replicaset-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:45 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-78fbc95b7 to 2 deployment-controller logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:46 +0000 UTC Normal Pod simple-prod-collector-5bf6cc7bcb-9s9bg AddedInterface Add eth0 [10.129.0.64/23] from ovn-kubernetes logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:46 +0000 UTC Normal Pod simple-prod-collector-5bf6cc7bcb-9s9bg.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:46 +0000 UTC Normal Pod simple-prod-collector-5bf6cc7bcb-9s9bg.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:46 +0000 UTC Normal Pod simple-prod-collector-5bf6cc7bcb-9s9bg.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:46 +0000 UTC Normal Pod simple-prod-collector-5bf6cc7bcb-wkkzn AddedInterface Add eth0 [10.130.0.49/23] from ovn-kubernetes logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:46 +0000 UTC Normal Pod simple-prod-collector-5bf6cc7bcb-wkkzn.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:46 +0000 UTC Normal Pod simple-prod-collector-5bf6cc7bcb-wkkzn.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:46 +0000 UTC Normal Pod simple-prod-collector-5bf6cc7bcb-wkkzn.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:46 +0000 UTC Normal Pod simple-prod-query-78fbc95b7-fxx66 AddedInterface Add eth0 [10.128.0.42/23] from ovn-kubernetes logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:46 +0000 UTC Normal Pod simple-prod-query-78fbc95b7-fxx66.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:46 +0000 UTC Normal Pod simple-prod-query-78fbc95b7-fxx66.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:46 +0000 UTC Normal Pod simple-prod-query-78fbc95b7-fxx66.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:46 +0000 UTC Normal Pod simple-prod-query-78fbc95b7-fxx66.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:46 +0000 UTC Normal Pod simple-prod-query-78fbc95b7-fxx66.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:46 +0000 UTC Normal Pod simple-prod-query-78fbc95b7-fxx66.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:46 +0000 UTC Normal Pod simple-prod-query-78fbc95b7-fxx66.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:46 +0000 UTC Normal Pod simple-prod-query-78fbc95b7-fxx66.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:46 +0000 UTC Normal Pod simple-prod-query-78fbc95b7-fxx66.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:46 +0000 UTC Normal Pod simple-prod-query-78fbc95b7-jchz5 AddedInterface Add eth0 [10.130.0.48/23] from ovn-kubernetes logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:46 +0000 UTC Normal Pod simple-prod-query-78fbc95b7-jchz5.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:46 +0000 UTC Normal Pod simple-prod-query-78fbc95b7-jchz5.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:46 +0000 UTC Normal Pod simple-prod-query-78fbc95b7-jchz5.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:46 +0000 UTC Normal Pod simple-prod-query-78fbc95b7-jchz5.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:46 +0000 UTC Normal Pod simple-prod-query-78fbc95b7-jchz5.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:46 +0000 UTC Normal Pod simple-prod-query-78fbc95b7-jchz5.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:46 +0000 UTC Normal Pod simple-prod-query-78fbc95b7-jchz5.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:46 +0000 UTC Normal Pod simple-prod-query-78fbc95b7-jchz5.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:11:50 | es-increasing-replicas | 2024-09-26 07:11:46 +0000 UTC Normal Pod simple-prod-query-78fbc95b7-jchz5.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:11:50 | es-increasing-replicas | Deleting namespace: kuttl-test-well-unicorn === CONT kuttl/harness/es-index-cleaner-autoprov logger.go:42: 07:12:29 | es-index-cleaner-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:12:29 | es-index-cleaner-autoprov | Creating namespace: kuttl-test-master-herring logger.go:42: 07:12:29 | es-index-cleaner-autoprov/1-install | starting test step 1-install logger.go:42: 07:12:29 | es-index-cleaner-autoprov/1-install | Jaeger:kuttl-test-master-herring/test-es-index-cleaner-with-prefix created logger.go:42: 07:13:04 | es-index-cleaner-autoprov/1-install | test step completed 1-install logger.go:42: 07:13:04 | es-index-cleaner-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 07:13:04 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE test-es-index-cleaner-with-prefix /dev/null] logger.go:42: 07:13:05 | es-index-cleaner-autoprov/2-report-spans | Warning: resource jaegers/test-es-index-cleaner-with-prefix is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:13:13 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c DAYS=5 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 07:13:13 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 07:13:14 | es-index-cleaner-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 07:13:55 | es-index-cleaner-autoprov/2-report-spans | test step completed 2-report-spans logger.go:42: 07:13:55 | es-index-cleaner-autoprov/3-install | starting test step 3-install logger.go:42: 07:13:56 | es-index-cleaner-autoprov/3-install | Jaeger:kuttl-test-master-herring/test-es-index-cleaner-with-prefix updated logger.go:42: 07:13:56 | es-index-cleaner-autoprov/3-install | test step completed 3-install logger.go:42: 07:13:56 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | starting test step 4-wait-es-index-cleaner logger.go:42: 07:13:56 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | running command: [sh -c go run ../../../../cmd-utils/wait-cronjob/main.go --cronjob test-es-index-cleaner-with-prefix-es-index-cleaner --namespace $NAMESPACE] logger.go:42: 07:13:57 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-09-26T07:13:57Z" level=debug msg="Checking if the test-es-index-cleaner-with-prefix-es-index-cleaner CronJob exists" logger.go:42: 07:13:57 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-09-26T07:13:57Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 07:13:57 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-09-26T07:13:57Z" level=info msg="Cronjob test-es-index-cleaner-with-prefix-es-index-cleaner found successfully" logger.go:42: 07:13:57 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-09-26T07:13:57Z" level=debug msg="Waiting for the next scheduled job from test-es-index-cleaner-with-prefix-es-index-cleaner cronjob" logger.go:42: 07:13:57 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-09-26T07:13:57Z" level=debug msg="Waiting for next job from test-es-index-cleaner-with-prefix-es-index-cleaner to succeed" logger.go:42: 07:14:07 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-09-26T07:14:07Z" level=info msg="Job of owner test-es-index-cleaner-with-prefix-es-index-cleaner succeeded after test-es-index-cleaner-with-prefix-es-index-cleaner 10.116775901s" logger.go:42: 07:14:07 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | test step completed 4-wait-es-index-cleaner logger.go:42: 07:14:07 | es-index-cleaner-autoprov/5-install | starting test step 5-install logger.go:42: 07:14:08 | es-index-cleaner-autoprov/5-install | Jaeger:kuttl-test-master-herring/test-es-index-cleaner-with-prefix updated logger.go:42: 07:14:08 | es-index-cleaner-autoprov/5-install | test step completed 5-install logger.go:42: 07:14:08 | es-index-cleaner-autoprov/6-check-indices | starting test step 6-check-indices logger.go:42: 07:14:08 | es-index-cleaner-autoprov/6-check-indices | Job:kuttl-test-master-herring/00-check-indices created logger.go:42: 07:14:13 | es-index-cleaner-autoprov/6-check-indices | test step completed 6-check-indices logger.go:42: 07:14:13 | es-index-cleaner-autoprov | es-index-cleaner-autoprov events from ns kuttl-test-master-herring: logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:12:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmasterherringtestesindexcle-1-76vffg Binding Scheduled Successfully assigned kuttl-test-master-herring/elasticsearch-cdm-kuttltestmasterherringtestesindexcle-1-76vffg to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:12:33 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmasterherringtestesindexcle-1-76vffg FailedMount MountVolume.SetUp failed for volume "elasticsearch-metrics" : secret "elasticsearch-metrics" not found kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:12:33 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestmasterherringtestesindexcle-1-7bd5bc47c5 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestmasterherringtestesindexcle-1-76vffg replicaset-controller logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:12:33 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestmasterherringtestesindexcle-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestmasterherringtestesindexcle-1-7bd5bc47c5 to 1 deployment-controller logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:12:34 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmasterherringtestesindexcle-1-76vffg AddedInterface Add eth0 [10.129.0.65/23] from ovn-kubernetes logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:12:34 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmasterherringtestesindexcle-1-76vffg.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:bc97b8e13087050dfb1a75b02c5b780cbb8fb12a849a655169d072ba8bbf42b4" already present on machine kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:12:34 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmasterherringtestesindexcle-1-76vffg.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:12:34 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmasterherringtestesindexcle-1-76vffg.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:12:34 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmasterherringtestesindexcle-1-76vffg.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:1d1d7d47b616995d18692f3a2d6232d7e1a0f41bc1503d87100f31939e080a78" already present on machine kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:12:34 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmasterherringtestesindexcle-1-76vffg.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:12:34 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmasterherringtestesindexcle-1-76vffg.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:12:44 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmasterherringtestesindexcle-1-76vffg.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:12:49 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmasterherringtestesindexcle-1-76vffg.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-7d69bd5c4b-sgmm6 Binding Scheduled Successfully assigned kuttl-test-master-herring/test-es-index-cleaner-with-prefix-collector-7d69bd5c4b-sgmm6 to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:00 +0000 UTC Warning Pod test-es-index-cleaner-with-prefix-collector-7d69bd5c4b-sgmm6 FailedMount MountVolume.SetUp failed for volume "test-es-index-cleaner-with-prefix-collector-tls-config-volume" : secret "test-es-index-cleaner-with-prefix-collector-headless-tls" not found kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:00 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-collector-7d69bd5c4b SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-collector-7d69bd5c4b-sgmm6 replicaset-controller logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:00 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-collector ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-collector-7d69bd5c4b to 1 deployment-controller logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-768bcdd7c6-dsvcr Binding Scheduled Successfully assigned kuttl-test-master-herring/test-es-index-cleaner-with-prefix-query-768bcdd7c6-dsvcr to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:00 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-768bcdd7c6 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-query-768bcdd7c6-dsvcr replicaset-controller logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:00 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-query-768bcdd7c6 to 1 deployment-controller logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:01 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-7d69bd5c4b-sgmm6 AddedInterface Add eth0 [10.130.0.50/23] from ovn-kubernetes logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:01 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-7d69bd5c4b-sgmm6.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:01 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-7d69bd5c4b-sgmm6.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:01 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-7d69bd5c4b-sgmm6.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:01 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-768bcdd7c6-dsvcr AddedInterface Add eth0 [10.128.0.43/23] from ovn-kubernetes logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:01 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-768bcdd7c6-dsvcr.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:01 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-768bcdd7c6-dsvcr.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:01 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-768bcdd7c6-dsvcr.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:01 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-768bcdd7c6-dsvcr.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:01 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-768bcdd7c6-dsvcr.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:01 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-768bcdd7c6-dsvcr.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:01 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-768bcdd7c6-dsvcr.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:01 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-768bcdd7c6-dsvcr.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:01 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-768bcdd7c6-dsvcr.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:06 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-768bcdd7c6-dsvcr.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:06 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-768bcdd7c6-dsvcr.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:06 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-768bcdd7c6-dsvcr.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:06 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-768bcdd7c6 SuccessfulDelete Deleted pod: test-es-index-cleaner-with-prefix-query-768bcdd7c6-dsvcr replicaset-controller logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:06 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled down replica set test-es-index-cleaner-with-prefix-query-768bcdd7c6 to 0 from 1 deployment-controller logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:08 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-594dccc5db-xf4jf Binding Scheduled Successfully assigned kuttl-test-master-herring/test-es-index-cleaner-with-prefix-query-594dccc5db-xf4jf to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:08 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-594dccc5db-xf4jf AddedInterface Add eth0 [10.128.0.44/23] from ovn-kubernetes logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:08 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-594dccc5db-xf4jf.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:08 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-594dccc5db-xf4jf.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:08 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-594dccc5db-xf4jf.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:08 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-594dccc5db-xf4jf.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:08 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-594dccc5db-xf4jf.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:08 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-594dccc5db-xf4jf.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:08 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-594dccc5db-xf4jf.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:08 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-594dccc5db-xf4jf.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:08 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-594dccc5db-xf4jf.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:08 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-594dccc5db SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-query-594dccc5db-xf4jf replicaset-controller logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:08 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-query-594dccc5db to 1 deployment-controller logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:14 +0000 UTC Normal Pod 00-report-span-l9kw2 Binding Scheduled Successfully assigned kuttl-test-master-herring/00-report-span-l9kw2 to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:14 +0000 UTC Normal Pod 00-report-span-l9kw2 AddedInterface Add eth0 [10.130.0.51/23] from ovn-kubernetes logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:14 +0000 UTC Normal Pod 00-report-span-l9kw2.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:14 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-l9kw2 job-controller logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:15 +0000 UTC Normal Pod 00-report-span-l9kw2.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 878ms (878ms including waiting) kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:15 +0000 UTC Normal Pod 00-report-span-l9kw2.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:15 +0000 UTC Normal Pod 00-report-span-l9kw2.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:15 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:15 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:15 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:45 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:45 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-7d69bd5c4b-sgmm6 horizontal-pod-autoscaler logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:45 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:13:54 +0000 UTC Normal Job.batch 00-report-span Completed Job completed job-controller logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:14:00 +0000 UTC Normal Job.batch test-es-index-cleaner-with-prefix-es-index-cleaner-28788914 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-es-index-cleaner-2878891x4vq5 job-controller logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:14:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2878891x4vq5 Binding Scheduled Successfully assigned kuttl-test-master-herring/test-es-index-cleaner-with-prefix-es-index-cleaner-2878891x4vq5 to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:14:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2878891x4vq5 AddedInterface Add eth0 [10.130.0.52/23] from ovn-kubernetes logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:14:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2878891x4vq5.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-index-cleaner-rhel8@sha256:503ae309c9a5dfd756ec3dc336874d6b7a37d1c3c061d771b6e7ee1b5a1e7912" kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:14:00 +0000 UTC Normal CronJob.batch test-es-index-cleaner-with-prefix-es-index-cleaner SuccessfulCreate Created job test-es-index-cleaner-with-prefix-es-index-cleaner-28788914 cronjob-controller logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:14:03 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2878891x4vq5.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-index-cleaner-rhel8@sha256:503ae309c9a5dfd756ec3dc336874d6b7a37d1c3c061d771b6e7ee1b5a1e7912" in 2.734s (2.734s including waiting) kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:14:03 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2878891x4vq5.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Created Created container test-es-index-cleaner-with-prefix-es-index-cleaner kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:14:03 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2878891x4vq5.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Started Started container test-es-index-cleaner-with-prefix-es-index-cleaner kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:14:06 +0000 UTC Normal Job.batch test-es-index-cleaner-with-prefix-es-index-cleaner-28788914 Completed Job completed job-controller logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:14:06 +0000 UTC Normal CronJob.batch test-es-index-cleaner-with-prefix-es-index-cleaner SawCompletedJob Saw completed job: test-es-index-cleaner-with-prefix-es-index-cleaner-28788914, status: Complete cronjob-controller logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:14:08 +0000 UTC Normal Pod 00-check-indices-7t9j9 Binding Scheduled Successfully assigned kuttl-test-master-herring/00-check-indices-7t9j9 to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:14:08 +0000 UTC Normal Pod 00-check-indices-7t9j9 AddedInterface Add eth0 [10.130.0.53/23] from ovn-kubernetes logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:14:08 +0000 UTC Normal Pod 00-check-indices-7t9j9.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:14:08 +0000 UTC Normal Job.batch 00-check-indices SuccessfulCreate Created pod: 00-check-indices-7t9j9 job-controller logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:14:09 +0000 UTC Normal Pod 00-check-indices-7t9j9.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 821ms (821ms including waiting) kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:14:09 +0000 UTC Normal Pod 00-check-indices-7t9j9.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:14:09 +0000 UTC Normal Pod 00-check-indices-7t9j9.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:14:13 | es-index-cleaner-autoprov | 2024-09-26 07:14:12 +0000 UTC Normal Job.batch 00-check-indices Completed Job completed job-controller logger.go:42: 07:14:14 | es-index-cleaner-autoprov | Deleting namespace: kuttl-test-master-herring === CONT kuttl/harness/es-from-aio-to-production logger.go:42: 07:14:20 | es-from-aio-to-production | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:14:20 | es-from-aio-to-production | Creating namespace: kuttl-test-fond-cardinal logger.go:42: 07:14:20 | es-from-aio-to-production/0-install | starting test step 0-install logger.go:42: 07:14:20 | es-from-aio-to-production/0-install | Jaeger:kuttl-test-fond-cardinal/my-jaeger created logger.go:42: 07:14:25 | es-from-aio-to-production/0-install | test step completed 0-install logger.go:42: 07:14:25 | es-from-aio-to-production/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:14:25 | es-from-aio-to-production/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:14:27 | es-from-aio-to-production/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:14:34 | es-from-aio-to-production/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:14:35 | es-from-aio-to-production/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:14:35 | es-from-aio-to-production/1-smoke-test | job.batch/report-span created logger.go:42: 07:14:35 | es-from-aio-to-production/1-smoke-test | job.batch/check-span created logger.go:42: 07:14:48 | es-from-aio-to-production/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:14:48 | es-from-aio-to-production/3-install | starting test step 3-install logger.go:42: 07:14:49 | es-from-aio-to-production/3-install | Jaeger:kuttl-test-fond-cardinal/my-jaeger updated logger.go:42: 07:15:21 | es-from-aio-to-production/3-install | test step completed 3-install logger.go:42: 07:15:21 | es-from-aio-to-production/4-smoke-test | starting test step 4-smoke-test logger.go:42: 07:15:21 | es-from-aio-to-production/4-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:15:30 | es-from-aio-to-production/4-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:15:30 | es-from-aio-to-production/4-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:15:31 | es-from-aio-to-production/4-smoke-test | job.batch/report-span unchanged logger.go:42: 07:15:31 | es-from-aio-to-production/4-smoke-test | job.batch/check-span unchanged logger.go:42: 07:15:31 | es-from-aio-to-production/4-smoke-test | test step completed 4-smoke-test logger.go:42: 07:15:31 | es-from-aio-to-production | es-from-aio-to-production events from ns kuttl-test-fond-cardinal: logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:21 +0000 UTC Normal Pod my-jaeger-75d4d6c46b-rzgwt Binding Scheduled Successfully assigned kuttl-test-fond-cardinal/my-jaeger-75d4d6c46b-rzgwt to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:21 +0000 UTC Normal Pod my-jaeger-75d4d6c46b-rzgwt AddedInterface Add eth0 [10.129.0.66/23] from ovn-kubernetes logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:21 +0000 UTC Normal Pod my-jaeger-75d4d6c46b-rzgwt.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:21 +0000 UTC Normal ReplicaSet.apps my-jaeger-75d4d6c46b SuccessfulCreate Created pod: my-jaeger-75d4d6c46b-rzgwt replicaset-controller logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:21 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-75d4d6c46b to 1 deployment-controller logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:24 +0000 UTC Normal Pod my-jaeger-75d4d6c46b-rzgwt.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" in 3.027s (3.027s including waiting) kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:24 +0000 UTC Normal Pod my-jaeger-75d4d6c46b-rzgwt.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:24 +0000 UTC Normal Pod my-jaeger-75d4d6c46b-rzgwt.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:24 +0000 UTC Normal Pod my-jaeger-75d4d6c46b-rzgwt.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:25 +0000 UTC Normal Pod my-jaeger-75d4d6c46b-rzgwt.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:25 +0000 UTC Normal Pod my-jaeger-75d4d6c46b-rzgwt.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:27 +0000 UTC Normal ReplicaSet.apps my-jaeger-75d4d6c46b SuccessfulDelete Deleted pod: my-jaeger-75d4d6c46b-rzgwt replicaset-controller logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:27 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-75d4d6c46b to 0 from 1 deployment-controller logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:28 +0000 UTC Normal Pod my-jaeger-75d4d6c46b-rzgwt.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:28 +0000 UTC Normal Pod my-jaeger-75d4d6c46b-rzgwt.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:29 +0000 UTC Normal Pod my-jaeger-5f47698665-shm59 Binding Scheduled Successfully assigned kuttl-test-fond-cardinal/my-jaeger-5f47698665-shm59 to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:29 +0000 UTC Normal Pod my-jaeger-5f47698665-shm59 AddedInterface Add eth0 [10.129.0.67/23] from ovn-kubernetes logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:29 +0000 UTC Normal Pod my-jaeger-5f47698665-shm59.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:29 +0000 UTC Normal Pod my-jaeger-5f47698665-shm59.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:29 +0000 UTC Normal Pod my-jaeger-5f47698665-shm59.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:29 +0000 UTC Normal Pod my-jaeger-5f47698665-shm59.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:29 +0000 UTC Normal Pod my-jaeger-5f47698665-shm59.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:29 +0000 UTC Normal Pod my-jaeger-5f47698665-shm59.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:29 +0000 UTC Normal ReplicaSet.apps my-jaeger-5f47698665 SuccessfulCreate Created pod: my-jaeger-5f47698665-shm59 replicaset-controller logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:29 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-5f47698665 to 1 deployment-controller logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:35 +0000 UTC Normal Pod check-span-bvvsw Binding Scheduled Successfully assigned kuttl-test-fond-cardinal/check-span-bvvsw to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:35 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-bvvsw job-controller logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:35 +0000 UTC Normal Pod report-span-2tt8b Binding Scheduled Successfully assigned kuttl-test-fond-cardinal/report-span-2tt8b to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:35 +0000 UTC Normal Pod report-span-2tt8b AddedInterface Add eth0 [10.128.0.45/23] from ovn-kubernetes logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:35 +0000 UTC Normal Pod report-span-2tt8b.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:35 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-2tt8b job-controller logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:36 +0000 UTC Normal Pod check-span-bvvsw AddedInterface Add eth0 [10.129.0.68/23] from ovn-kubernetes logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:36 +0000 UTC Normal Pod check-span-bvvsw.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:37 +0000 UTC Normal Pod check-span-bvvsw.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 894ms (894ms including waiting) kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:37 +0000 UTC Normal Pod check-span-bvvsw.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:37 +0000 UTC Normal Pod check-span-bvvsw.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:37 +0000 UTC Normal Pod report-span-2tt8b.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.435s (1.435s including waiting) kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:37 +0000 UTC Normal Pod report-span-2tt8b.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:37 +0000 UTC Normal Pod report-span-2tt8b.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:48 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:51 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestfondcardinalmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestfondcardinalmyjaeger-1-6dcf9869b5 to 1 deployment-controller logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:52 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfondcardinalmyjaeger-1-6dcf9869878qz Binding Scheduled Successfully assigned kuttl-test-fond-cardinal/elasticsearch-cdm-kuttltestfondcardinalmyjaeger-1-6dcf9869878qz to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:52 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfondcardinalmyjaeger-1-6dcf9869878qz AddedInterface Add eth0 [10.129.0.69/23] from ovn-kubernetes logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:52 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfondcardinalmyjaeger-1-6dcf9869878qz.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:bc97b8e13087050dfb1a75b02c5b780cbb8fb12a849a655169d072ba8bbf42b4" already present on machine kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:52 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfondcardinalmyjaeger-1-6dcf9869878qz.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:52 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfondcardinalmyjaeger-1-6dcf9869878qz.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:52 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfondcardinalmyjaeger-1-6dcf9869878qz.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:1d1d7d47b616995d18692f3a2d6232d7e1a0f41bc1503d87100f31939e080a78" already present on machine kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:52 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfondcardinalmyjaeger-1-6dcf9869878qz.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:52 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfondcardinalmyjaeger-1-6dcf9869878qz.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:14:52 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestfondcardinalmyjaeger-1-6dcf9869b5 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestfondcardinalmyjaeger-1-6dcf9869878qz replicaset-controller logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:15:02 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestfondcardinalmyjaeger-1-6dcf9869878qz.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:15:07 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestfondcardinalmyjaeger-1-6dcf9869878qz.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:15:11 +0000 UTC Normal Job.batch report-span Completed Job completed job-controller logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:15:18 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-644b59884 SuccessfulCreate Created pod: my-jaeger-collector-644b59884-m942z replicaset-controller logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:15:18 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-644b59884 to 1 deployment-controller logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:15:18 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-76d9788595 to 1 deployment-controller logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:15:19 +0000 UTC Normal Pod my-jaeger-collector-644b59884-m942z Binding Scheduled Successfully assigned kuttl-test-fond-cardinal/my-jaeger-collector-644b59884-m942z to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:15:19 +0000 UTC Normal Pod my-jaeger-collector-644b59884-m942z AddedInterface Add eth0 [10.130.0.54/23] from ovn-kubernetes logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:15:19 +0000 UTC Normal Pod my-jaeger-collector-644b59884-m942z.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:15:19 +0000 UTC Normal Pod my-jaeger-collector-644b59884-m942z.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:15:19 +0000 UTC Normal Pod my-jaeger-collector-644b59884-m942z.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:15:19 +0000 UTC Normal Pod my-jaeger-query-76d9788595-wncrd Binding Scheduled Successfully assigned kuttl-test-fond-cardinal/my-jaeger-query-76d9788595-wncrd to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:15:19 +0000 UTC Normal Pod my-jaeger-query-76d9788595-wncrd AddedInterface Add eth0 [10.130.0.55/23] from ovn-kubernetes logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:15:19 +0000 UTC Normal Pod my-jaeger-query-76d9788595-wncrd.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:15:19 +0000 UTC Normal Pod my-jaeger-query-76d9788595-wncrd.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:15:19 +0000 UTC Normal Pod my-jaeger-query-76d9788595-wncrd.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:15:19 +0000 UTC Normal Pod my-jaeger-query-76d9788595-wncrd.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:15:19 +0000 UTC Normal Pod my-jaeger-query-76d9788595-wncrd.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:15:19 +0000 UTC Normal Pod my-jaeger-query-76d9788595-wncrd.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:15:19 +0000 UTC Normal Pod my-jaeger-query-76d9788595-wncrd.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:15:19 +0000 UTC Normal Pod my-jaeger-query-76d9788595-wncrd.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:15:19 +0000 UTC Normal Pod my-jaeger-query-76d9788595-wncrd.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:15:19 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-76d9788595 SuccessfulCreate Created pod: my-jaeger-query-76d9788595-wncrd replicaset-controller logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:15:22 +0000 UTC Normal Pod my-jaeger-5f47698665-shm59.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | 2024-09-26 07:15:22 +0000 UTC Normal Pod my-jaeger-5f47698665-shm59.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:15:31 | es-from-aio-to-production | Deleting namespace: kuttl-test-fond-cardinal === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (682.02s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.95s) --- PASS: kuttl/harness/es-multiinstance (81.21s) --- PASS: kuttl/harness/es-simple-prod (5.99s) --- PASS: kuttl/harness/es-rollover-autoprov (275.04s) --- PASS: kuttl/harness/es-increasing-replicas (124.05s) --- PASS: kuttl/harness/es-index-cleaner-autoprov (111.36s) --- PASS: kuttl/harness/es-from-aio-to-production (77.82s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name elasticsearch --report --output /logs/artifacts/elasticsearch.xml ./artifacts/kuttl-report.xml time="2024-09-26T07:15:38Z" level=debug msg="Setting a new name for the test suites" time="2024-09-26T07:15:38Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-09-26T07:15:38Z" level=debug msg="normalizing test case names" time="2024-09-26T07:15:38Z" level=debug msg="elasticsearch/artifacts -> elasticsearch_artifacts" time="2024-09-26T07:15:38Z" level=debug msg="elasticsearch/es-multiinstance -> elasticsearch_es_multiinstance" time="2024-09-26T07:15:38Z" level=debug msg="elasticsearch/es-simple-prod -> elasticsearch_es_simple_prod" time="2024-09-26T07:15:38Z" level=debug msg="elasticsearch/es-rollover-autoprov -> elasticsearch_es_rollover_autoprov" time="2024-09-26T07:15:38Z" level=debug msg="elasticsearch/es-increasing-replicas -> elasticsearch_es_increasing_replicas" time="2024-09-26T07:15:38Z" level=debug msg="elasticsearch/es-index-cleaner-autoprov -> elasticsearch_es_index_cleaner_autoprov" time="2024-09-26T07:15:38Z" level=debug msg="elasticsearch/es-from-aio-to-production -> elasticsearch_es_from_aio_to_production" +-----------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------+--------+ | elasticsearch_artifacts | passed | | elasticsearch_es_multiinstance | passed | | elasticsearch_es_simple_prod | passed | | elasticsearch_es_rollover_autoprov | passed | | elasticsearch_es_increasing_replicas | passed | | elasticsearch_es_index_cleaner_autoprov | passed | | elasticsearch_es_from_aio_to_production | passed | +-----------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh examples false true + '[' 3 -ne 3 ']' + test_suite_name=examples + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/examples.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-examples make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ VERTX_IMG=jaegertracing/vertx-create-span:operator-e2e-tests \ ./tests/e2e/examples/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.32 True False 29m Cluster version is 4.15.32' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.32 True False 29m Cluster version is 4.15.32' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/examples/render.sh ++ export SUITE_DIR=./tests/e2e/examples ++ SUITE_DIR=./tests/e2e/examples ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/examples ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test examples-agent-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-agent-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-with-priority-class\e[0m' Rendering files for test examples-agent-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + mkdir -p examples-agent-with-priority-class + cd examples-agent-with-priority-class + example_name=agent-with-priority-class + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + '[' true '!=' true ']' + render_install_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + install_file=./02-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/agent-with-priority-class.yaml -o ./02-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./02-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./02-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./02-install.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + JAEGER_NAME=agent-as-daemonset + local jaeger_strategy ++ get_jaeger_strategy ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./02-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./02-install.yaml ++ strategy=DaemonSet ++ '[' DaemonSet = null ']' ++ echo DaemonSet ++ return 0 + jaeger_strategy=DaemonSet + '[' DaemonSet = DaemonSet ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./02-assert.yaml + render_smoke_test_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + jaeger_name=agent-as-daemonset + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test agent-as-daemonset true 02 + '[' 3 -ne 3 ']' + jaeger=agent-as-daemonset + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + export JAEGER_NAME=agent-as-daemonset + JAEGER_NAME=agent-as-daemonset + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-all-in-one-with-options + '[' 1 -ne 1 ']' + test_name=examples-all-in-one-with-options + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-all-in-one-with-options' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-all-in-one-with-options\e[0m' Rendering files for test examples-all-in-one-with-options + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-with-priority-class + '[' examples-agent-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-all-in-one-with-options + cd examples-all-in-one-with-options + example_name=all-in-one-with-options + render_install_example all-in-one-with-options 00 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/all-in-one-with-options.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + JAEGER_NAME=my-jaeger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.metadata.name="my-jaeger"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i 'del(.spec.allInOne.image)' ./00-install.yaml + render_smoke_test_example all-in-one-with-options 01 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + jaeger_name=my-jaeger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + sed -i s~my-jaeger-query:443~my-jaeger-query:443/jaeger~gi ./01-smoke-test.yaml + start_test examples-business-application-injected-sidecar + '[' 1 -ne 1 ']' + test_name=examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-business-application-injected-sidecar' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-business-application-injected-sidecar\e[0m' Rendering files for test examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-all-in-one-with-options + '[' examples-all-in-one-with-options '!=' _build ']' + cd .. + mkdir -p examples-business-application-injected-sidecar + cd examples-business-application-injected-sidecar + example_name=simplest + cp /tmp/jaeger-tests/examples/business-application-injected-sidecar.yaml ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].image=strenv(VERTX_IMG)' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.path="/"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.port=8080' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.initialDelaySeconds=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.failureThreshold=3' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.periodSeconds=10' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.successThreshold=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.timeoutSeconds=1' ./00-install.yaml + render_install_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example simplest 02 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 02 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-collector-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-collector-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-collector-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-collector-with-priority-class\e[0m' Rendering files for test examples-collector-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-business-application-injected-sidecar + '[' examples-business-application-injected-sidecar '!=' _build ']' + cd .. + mkdir -p examples-collector-with-priority-class + cd examples-collector-with-priority-class + example_name=collector-with-priority-class + render_install_example collector-with-priority-class 00 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/collector-with-priority-class.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + JAEGER_NAME=collector-with-high-priority + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example collector-with-priority-class 01 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + jaeger_name=collector-with-high-priority + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test collector-with-high-priority true 01 + '[' 3 -ne 3 ']' + jaeger=collector-with-high-priority + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + export JAEGER_NAME=collector-with-high-priority + JAEGER_NAME=collector-with-high-priority + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-service-types + '[' 1 -ne 1 ']' + test_name=examples-service-types + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-service-types' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-service-types\e[0m' Rendering files for test examples-service-types + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-collector-with-priority-class + '[' examples-collector-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-service-types + cd examples-service-types + example_name=service-types + render_install_example service-types 00 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/service-types.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + JAEGER_NAME=service-types + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example service-types 01 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/service-types.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/service-types.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/service-types.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/service-types.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + jaeger_name=service-types + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test service-types true 01 + '[' 3 -ne 3 ']' + jaeger=service-types + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + export JAEGER_NAME=service-types + JAEGER_NAME=service-types + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod + '[' 1 -ne 1 ']' + test_name=examples-simple-prod + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod\e[0m' Rendering files for test examples-simple-prod + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-service-types + '[' examples-service-types '!=' _build ']' + cd .. + mkdir -p examples-simple-prod + cd examples-simple-prod + example_name=simple-prod + render_install_example simple-prod 01 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod 02 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod-with-volumes + '[' 1 -ne 1 ']' + test_name=examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod-with-volumes' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod-with-volumes\e[0m' Rendering files for test examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod + '[' examples-simple-prod '!=' _build ']' + cd .. + mkdir -p examples-simple-prod-with-volumes + cd examples-simple-prod-with-volumes + example_name=simple-prod-with-volumes + render_install_example simple-prod-with-volumes 01 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod-with-volumes 02 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + /tmp/jaeger-tests/bin/gomplate -f ./03-check-volume.yaml.template -o 03-check-volume.yaml + start_test examples-simplest + '[' 1 -ne 1 ']' + test_name=examples-simplest + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simplest' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simplest\e[0m' Rendering files for test examples-simplest + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod-with-volumes + '[' examples-simple-prod-with-volumes '!=' _build ']' + cd .. + mkdir -p examples-simplest + cd examples-simplest + example_name=simplest + render_install_example simplest 00 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 01 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger + '[' 1 -ne 1 ']' + test_name=examples-with-badger + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger\e[0m' Rendering files for test examples-with-badger + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simplest + '[' examples-simplest '!=' _build ']' + cd .. + mkdir -p examples-with-badger + cd examples-with-badger + example_name=with-badger + render_install_example with-badger 00 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + JAEGER_NAME=with-badger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger 01 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + jaeger_name=with-badger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + export JAEGER_NAME=with-badger + JAEGER_NAME=with-badger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger-and-volume + '[' 1 -ne 1 ']' + test_name=examples-with-badger-and-volume + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger-and-volume' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger-and-volume\e[0m' Rendering files for test examples-with-badger-and-volume + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger + '[' examples-with-badger '!=' _build ']' + cd .. + mkdir -p examples-with-badger-and-volume + cd examples-with-badger-and-volume + example_name=with-badger-and-volume + render_install_example with-badger-and-volume 00 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger-and-volume.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + JAEGER_NAME=with-badger-and-volume + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger-and-volume 01 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + jaeger_name=with-badger-and-volume + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger-and-volume true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger-and-volume + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + export JAEGER_NAME=with-badger-and-volume + JAEGER_NAME=with-badger-and-volume + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-cassandra + '[' 1 -ne 1 ']' + test_name=examples-with-cassandra + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-cassandra' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-cassandra\e[0m' Rendering files for test examples-with-cassandra + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger-and-volume + '[' examples-with-badger-and-volume '!=' _build ']' + cd .. + mkdir -p examples-with-cassandra + cd examples-with-cassandra + example_name=with-cassandra + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-cassandra 01 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-cassandra.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + JAEGER_NAME=with-cassandra + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-cassandra 02 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-cassandra.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-cassandra.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + jaeger_name=with-cassandra + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-cassandra true 02 + '[' 3 -ne 3 ']' + jaeger=with-cassandra + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + export JAEGER_NAME=with-cassandra + JAEGER_NAME=with-cassandra + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-sampling + '[' 1 -ne 1 ']' + test_name=examples-with-sampling + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-sampling' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-sampling\e[0m' Rendering files for test examples-with-sampling + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-cassandra + '[' examples-with-cassandra '!=' _build ']' + cd .. + mkdir -p examples-with-sampling + cd examples-with-sampling + export example_name=with-sampling + example_name=with-sampling + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-sampling 01 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-sampling.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + JAEGER_NAME=with-sampling + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-sampling 02 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-sampling.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-sampling.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + jaeger_name=with-sampling + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-sampling true 02 + '[' 3 -ne 3 ']' + jaeger=with-sampling + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + export JAEGER_NAME=with-sampling + JAEGER_NAME=with-sampling + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-agent-as-daemonset + '[' 1 -ne 1 ']' + test_name=examples-agent-as-daemonset + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-as-daemonset' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-as-daemonset\e[0m' Rendering files for test examples-agent-as-daemonset + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-sampling + '[' examples-with-sampling '!=' _build ']' + cd .. + mkdir -p examples-agent-as-daemonset + cd examples-agent-as-daemonset + '[' true = true ']' + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/agent-as-daemonset.yaml -o 02-install.yaml + '[' true = true ']' + start_test examples-openshift-with-htpasswd + '[' 1 -ne 1 ']' + test_name=examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-openshift-with-htpasswd' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-openshift-with-htpasswd\e[0m' Rendering files for test examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-as-daemonset + '[' examples-agent-as-daemonset '!=' _build ']' + cd .. + mkdir -p examples-openshift-with-htpasswd + cd examples-openshift-with-htpasswd + export JAEGER_NAME=with-htpasswd + JAEGER_NAME=with-htpasswd + export JAEGER_USERNAME=awesomeuser + JAEGER_USERNAME=awesomeuser + export JAEGER_PASSWORD=awesomepassword + JAEGER_PASSWORD=awesomepassword + export 'JAEGER_USER_PASSWORD_HASH=awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' + JAEGER_USER_PASSWORD_HASH='awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ echo 'awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ base64 + SECRET=YXdlc29tZXVzZXI6e1NIQX11VWRxUFZVeXFOQm1FUlUwUXhqM0tGYVpuanc9Cg== + /tmp/jaeger-tests/bin/gomplate -f ./00-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/with-htpasswd.yaml -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + export 'GET_URL_COMMAND=kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + export 'URL=https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + INSECURE=true + JAEGER_USERNAME= + JAEGER_PASSWORD= + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-unsecured.yaml + JAEGER_USERNAME=wronguser + JAEGER_PASSWORD=wrongpassword + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-check-unauthorized.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./04-check-authorized.yaml + skip_test examples-agent-as-daemonset 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-agent-as-daemonset + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-openshift-with-htpasswd + '[' examples-openshift-with-htpasswd '!=' _build ']' + cd .. + rm -rf examples-agent-as-daemonset + warning 'examples-agent-as-daemonset: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-agent-as-daemonset: This test is flaky in Prow CI\e[0m' WAR: examples-agent-as-daemonset: This test is flaky in Prow CI + skip_test examples-with-badger-and-volume 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-with-badger-and-volume + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + rm -rf examples-with-badger-and-volume + warning 'examples-with-badger-and-volume: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-with-badger-and-volume: This test is flaky in Prow CI\e[0m' WAR: examples-with-badger-and-volume: This test is flaky in Prow CI + skip_test examples-collector-with-priority-class 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-collector-with-priority-class + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + rm -rf examples-collector-with-priority-class + warning 'examples-collector-with-priority-class: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-collector-with-priority-class: This test is flaky in Prow CI\e[0m' WAR: examples-collector-with-priority-class: This test is flaky in Prow CI make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running examples E2E tests' Running examples E2E tests + cd tests/e2e/examples/_build + set +e + KUBECONFIG=/tmp/kubeconfig-488946273 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-rosa-h-ezi9.l97m.s3.devshift.org:443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 12 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/examples-agent-with-priority-class === PAUSE kuttl/harness/examples-agent-with-priority-class === RUN kuttl/harness/examples-all-in-one-with-options === PAUSE kuttl/harness/examples-all-in-one-with-options === RUN kuttl/harness/examples-business-application-injected-sidecar === PAUSE kuttl/harness/examples-business-application-injected-sidecar === RUN kuttl/harness/examples-openshift-with-htpasswd === PAUSE kuttl/harness/examples-openshift-with-htpasswd === RUN kuttl/harness/examples-service-types === PAUSE kuttl/harness/examples-service-types === RUN kuttl/harness/examples-simple-prod === PAUSE kuttl/harness/examples-simple-prod === RUN kuttl/harness/examples-simple-prod-with-volumes === PAUSE kuttl/harness/examples-simple-prod-with-volumes === RUN kuttl/harness/examples-simplest === PAUSE kuttl/harness/examples-simplest === RUN kuttl/harness/examples-with-badger === PAUSE kuttl/harness/examples-with-badger === RUN kuttl/harness/examples-with-cassandra === PAUSE kuttl/harness/examples-with-cassandra === RUN kuttl/harness/examples-with-sampling === PAUSE kuttl/harness/examples-with-sampling === CONT kuttl/harness/artifacts logger.go:42: 07:16:12 | artifacts | Creating namespace: kuttl-test-tolerant-buzzard logger.go:42: 07:16:12 | artifacts | artifacts events from ns kuttl-test-tolerant-buzzard: logger.go:42: 07:16:12 | artifacts | Deleting namespace: kuttl-test-tolerant-buzzard === CONT kuttl/harness/examples-simple-prod logger.go:42: 07:16:18 | examples-simple-prod | Creating namespace: kuttl-test-fond-imp logger.go:42: 07:16:18 | examples-simple-prod/1-install | starting test step 1-install logger.go:42: 07:16:18 | examples-simple-prod/1-install | Jaeger:kuttl-test-fond-imp/simple-prod created logger.go:42: 07:16:51 | examples-simple-prod/1-install | test step completed 1-install logger.go:42: 07:16:51 | examples-simple-prod/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:16:51 | examples-simple-prod/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:16:53 | examples-simple-prod/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:17:01 | examples-simple-prod/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:17:01 | examples-simple-prod/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:17:01 | examples-simple-prod/2-smoke-test | job.batch/report-span created logger.go:42: 07:17:02 | examples-simple-prod/2-smoke-test | job.batch/check-span created logger.go:42: 07:17:15 | examples-simple-prod/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:17:15 | examples-simple-prod | examples-simple-prod events from ns kuttl-test-fond-imp: logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:22 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfondimpsimpleprod-1-74565766d5-22k5p Binding Scheduled Successfully assigned kuttl-test-fond-imp/elasticsearch-cdm-kuttltestfondimpsimpleprod-1-74565766d5-22k5p to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:22 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestfondimpsimpleprod-1-74565766d5 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestfondimpsimpleprod-1-74565766d5-22k5p replicaset-controller logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:22 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestfondimpsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestfondimpsimpleprod-1-74565766d5 to 1 deployment-controller logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:23 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfondimpsimpleprod-1-74565766d5-22k5p AddedInterface Add eth0 [10.129.0.70/23] from ovn-kubernetes logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:23 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfondimpsimpleprod-1-74565766d5-22k5p.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:bc97b8e13087050dfb1a75b02c5b780cbb8fb12a849a655169d072ba8bbf42b4" already present on machine kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:23 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfondimpsimpleprod-1-74565766d5-22k5p.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:23 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfondimpsimpleprod-1-74565766d5-22k5p.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:23 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfondimpsimpleprod-1-74565766d5-22k5p.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:1d1d7d47b616995d18692f3a2d6232d7e1a0f41bc1503d87100f31939e080a78" already present on machine kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:23 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfondimpsimpleprod-1-74565766d5-22k5p.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:23 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfondimpsimpleprod-1-74565766d5-22k5p.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:33 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestfondimpsimpleprod-1-74565766d5-22k5p.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:38 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestfondimpsimpleprod-1-74565766d5-22k5p.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:49 +0000 UTC Normal Pod simple-prod-collector-6f55fd4df4-q4wl6 Binding Scheduled Successfully assigned kuttl-test-fond-imp/simple-prod-collector-6f55fd4df4-q4wl6 to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:49 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-6f55fd4df4 SuccessfulCreate Created pod: simple-prod-collector-6f55fd4df4-q4wl6 replicaset-controller logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:49 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-6f55fd4df4 to 1 deployment-controller logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:49 +0000 UTC Normal Pod simple-prod-query-9d46c7796-fxwlt Binding Scheduled Successfully assigned kuttl-test-fond-imp/simple-prod-query-9d46c7796-fxwlt to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:49 +0000 UTC Normal ReplicaSet.apps simple-prod-query-9d46c7796 SuccessfulCreate Created pod: simple-prod-query-9d46c7796-fxwlt replicaset-controller logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:49 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-9d46c7796 to 1 deployment-controller logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:50 +0000 UTC Normal Pod simple-prod-collector-6f55fd4df4-q4wl6 AddedInterface Add eth0 [10.130.0.56/23] from ovn-kubernetes logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:50 +0000 UTC Normal Pod simple-prod-collector-6f55fd4df4-q4wl6.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:50 +0000 UTC Normal Pod simple-prod-collector-6f55fd4df4-q4wl6.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:50 +0000 UTC Normal Pod simple-prod-collector-6f55fd4df4-q4wl6.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:50 +0000 UTC Normal Pod simple-prod-query-9d46c7796-fxwlt AddedInterface Add eth0 [10.128.0.46/23] from ovn-kubernetes logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:50 +0000 UTC Normal Pod simple-prod-query-9d46c7796-fxwlt.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:50 +0000 UTC Normal Pod simple-prod-query-9d46c7796-fxwlt.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:50 +0000 UTC Normal Pod simple-prod-query-9d46c7796-fxwlt.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:50 +0000 UTC Normal Pod simple-prod-query-9d46c7796-fxwlt.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:50 +0000 UTC Normal Pod simple-prod-query-9d46c7796-fxwlt.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:50 +0000 UTC Normal Pod simple-prod-query-9d46c7796-fxwlt.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:50 +0000 UTC Normal Pod simple-prod-query-9d46c7796-fxwlt.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:50 +0000 UTC Normal Pod simple-prod-query-9d46c7796-fxwlt.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:50 +0000 UTC Normal Pod simple-prod-query-9d46c7796-fxwlt.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:54 +0000 UTC Normal Pod simple-prod-query-9d46c7796-fxwlt.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:54 +0000 UTC Normal Pod simple-prod-query-9d46c7796-fxwlt.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:54 +0000 UTC Normal Pod simple-prod-query-9d46c7796-fxwlt.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:54 +0000 UTC Normal ReplicaSet.apps simple-prod-query-9d46c7796 SuccessfulDelete Deleted pod: simple-prod-query-9d46c7796-fxwlt replicaset-controller logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:54 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-9d46c7796 to 0 from 1 deployment-controller logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:55 +0000 UTC Normal Pod simple-prod-query-58f9d597b4-lrhz9 Binding Scheduled Successfully assigned kuttl-test-fond-imp/simple-prod-query-58f9d597b4-lrhz9 to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:55 +0000 UTC Normal ReplicaSet.apps simple-prod-query-58f9d597b4 SuccessfulCreate Created pod: simple-prod-query-58f9d597b4-lrhz9 replicaset-controller logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:55 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-58f9d597b4 to 1 deployment-controller logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:56 +0000 UTC Normal Pod simple-prod-query-58f9d597b4-lrhz9 AddedInterface Add eth0 [10.128.0.47/23] from ovn-kubernetes logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:56 +0000 UTC Normal Pod simple-prod-query-58f9d597b4-lrhz9.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:56 +0000 UTC Normal Pod simple-prod-query-58f9d597b4-lrhz9.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:56 +0000 UTC Normal Pod simple-prod-query-58f9d597b4-lrhz9.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:56 +0000 UTC Normal Pod simple-prod-query-58f9d597b4-lrhz9.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:56 +0000 UTC Normal Pod simple-prod-query-58f9d597b4-lrhz9.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:56 +0000 UTC Normal Pod simple-prod-query-58f9d597b4-lrhz9.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:56 +0000 UTC Normal Pod simple-prod-query-58f9d597b4-lrhz9.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:56 +0000 UTC Normal Pod simple-prod-query-58f9d597b4-lrhz9.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:16:56 +0000 UTC Normal Pod simple-prod-query-58f9d597b4-lrhz9.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:17:01 +0000 UTC Normal Pod report-span-hf29r Binding Scheduled Successfully assigned kuttl-test-fond-imp/report-span-hf29r to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:17:01 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-hf29r job-controller logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:17:02 +0000 UTC Normal Pod check-span-jsgsr Binding Scheduled Successfully assigned kuttl-test-fond-imp/check-span-jsgsr to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:17:02 +0000 UTC Normal Pod check-span-jsgsr AddedInterface Add eth0 [10.130.0.58/23] from ovn-kubernetes logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:17:02 +0000 UTC Normal Pod check-span-jsgsr.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:17:02 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-jsgsr job-controller logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:17:02 +0000 UTC Normal Pod report-span-hf29r AddedInterface Add eth0 [10.130.0.57/23] from ovn-kubernetes logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:17:02 +0000 UTC Normal Pod report-span-hf29r.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:17:03 +0000 UTC Normal Pod check-span-jsgsr.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 833ms (833ms including waiting) kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:17:03 +0000 UTC Normal Pod check-span-jsgsr.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:17:03 +0000 UTC Normal Pod check-span-jsgsr.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:17:03 +0000 UTC Normal Pod report-span-hf29r.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 850ms (850ms including waiting) kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:17:03 +0000 UTC Normal Pod report-span-hf29r.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:17:03 +0000 UTC Normal Pod report-span-hf29r.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:17:04 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:17:04 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:17:04 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:17:15 | examples-simple-prod | 2024-09-26 07:17:14 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:17:15 | examples-simple-prod | Deleting namespace: kuttl-test-fond-imp === CONT kuttl/harness/examples-with-sampling logger.go:42: 07:17:27 | examples-with-sampling | Creating namespace: kuttl-test-dynamic-camel logger.go:42: 07:17:27 | examples-with-sampling/0-install | starting test step 0-install logger.go:42: 07:17:27 | examples-with-sampling/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 07:17:27 | examples-with-sampling/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 07:17:27 | examples-with-sampling/0-install | >>>> Creating namespace kuttl-test-dynamic-camel logger.go:42: 07:17:27 | examples-with-sampling/0-install | kubectl create namespace kuttl-test-dynamic-camel 2>&1 | grep -v "already exists" || true logger.go:42: 07:17:27 | examples-with-sampling/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-dynamic-camel 2>&1 | grep -v "already exists" || true logger.go:42: 07:17:28 | examples-with-sampling/0-install | service/cassandra created logger.go:42: 07:17:28 | examples-with-sampling/0-install | statefulset.apps/cassandra created logger.go:42: 07:17:28 | examples-with-sampling/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 07:17:42 | examples-with-sampling/0-install | test step completed 0-install logger.go:42: 07:17:42 | examples-with-sampling/1-install | starting test step 1-install logger.go:42: 07:17:42 | examples-with-sampling/1-install | Jaeger:kuttl-test-dynamic-camel/with-sampling created logger.go:42: 07:17:44 | examples-with-sampling/1-install | test step completed 1-install logger.go:42: 07:17:44 | examples-with-sampling/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:17:44 | examples-with-sampling/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-sampling /dev/null] logger.go:42: 07:17:46 | examples-with-sampling/2-smoke-test | Warning: resource jaegers/with-sampling is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:17:53 | examples-with-sampling/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:17:54 | examples-with-sampling/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:17:54 | examples-with-sampling/2-smoke-test | job.batch/report-span created logger.go:42: 07:17:55 | examples-with-sampling/2-smoke-test | job.batch/check-span created logger.go:42: 07:18:07 | examples-with-sampling/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:18:07 | examples-with-sampling/3- | starting test step 3- logger.go:42: 07:18:07 | examples-with-sampling/3- | test step completed 3- logger.go:42: 07:18:07 | examples-with-sampling | examples-with-sampling events from ns kuttl-test-dynamic-camel: logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:28 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-dynamic-camel/cassandra-0 to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:28 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.128.0.48/23] from ovn-kubernetes logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:28 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:28 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:34 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 5.703s (5.703s including waiting) kubelet logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:34 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:34 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:34 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-dynamic-camel/cassandra-1 to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:34 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:35 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.129.0.71/23] from ovn-kubernetes logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:35 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:40 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 5.712s (5.712s including waiting) kubelet logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:41 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:41 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:43 +0000 UTC Normal Pod with-sampling-5f8487b468-vzrdn Binding Scheduled Successfully assigned kuttl-test-dynamic-camel/with-sampling-5f8487b468-vzrdn to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:43 +0000 UTC Normal Pod with-sampling-5f8487b468-vzrdn AddedInterface Add eth0 [10.129.0.72/23] from ovn-kubernetes logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:43 +0000 UTC Normal Pod with-sampling-5f8487b468-vzrdn.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:43 +0000 UTC Normal Pod with-sampling-5f8487b468-vzrdn.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:43 +0000 UTC Normal Pod with-sampling-5f8487b468-vzrdn.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:43 +0000 UTC Normal Pod with-sampling-5f8487b468-vzrdn.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:43 +0000 UTC Normal Pod with-sampling-5f8487b468-vzrdn.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:43 +0000 UTC Normal Pod with-sampling-5f8487b468-vzrdn.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:43 +0000 UTC Normal ReplicaSet.apps with-sampling-5f8487b468 SuccessfulCreate Created pod: with-sampling-5f8487b468-vzrdn replicaset-controller logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:43 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-5f8487b468 to 1 deployment-controller logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:47 +0000 UTC Normal Pod with-sampling-5f8487b468-vzrdn.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:47 +0000 UTC Normal Pod with-sampling-5f8487b468-vzrdn.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:47 +0000 UTC Normal ReplicaSet.apps with-sampling-5f8487b468 SuccessfulDelete Deleted pod: with-sampling-5f8487b468-vzrdn replicaset-controller logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:47 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled down replica set with-sampling-5f8487b468 to 0 from 1 deployment-controller logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:48 +0000 UTC Normal Pod with-sampling-6b48c465ff-9jb4d Binding Scheduled Successfully assigned kuttl-test-dynamic-camel/with-sampling-6b48c465ff-9jb4d to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:48 +0000 UTC Normal ReplicaSet.apps with-sampling-6b48c465ff SuccessfulCreate Created pod: with-sampling-6b48c465ff-9jb4d replicaset-controller logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:48 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-6b48c465ff to 1 deployment-controller logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:49 +0000 UTC Normal Pod with-sampling-6b48c465ff-9jb4d AddedInterface Add eth0 [10.129.0.73/23] from ovn-kubernetes logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:49 +0000 UTC Normal Pod with-sampling-6b48c465ff-9jb4d.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:49 +0000 UTC Normal Pod with-sampling-6b48c465ff-9jb4d.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:49 +0000 UTC Normal Pod with-sampling-6b48c465ff-9jb4d.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:49 +0000 UTC Normal Pod with-sampling-6b48c465ff-9jb4d.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:49 +0000 UTC Normal Pod with-sampling-6b48c465ff-9jb4d.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:49 +0000 UTC Normal Pod with-sampling-6b48c465ff-9jb4d.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:54 +0000 UTC Normal Pod report-span-q5l4g Binding Scheduled Successfully assigned kuttl-test-dynamic-camel/report-span-q5l4g to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:54 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-q5l4g job-controller logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:55 +0000 UTC Normal Pod check-span-twqhx Binding Scheduled Successfully assigned kuttl-test-dynamic-camel/check-span-twqhx to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:55 +0000 UTC Normal Pod check-span-twqhx AddedInterface Add eth0 [10.130.0.60/23] from ovn-kubernetes logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:55 +0000 UTC Normal Pod check-span-twqhx.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:55 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-twqhx job-controller logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:55 +0000 UTC Normal Pod report-span-q5l4g AddedInterface Add eth0 [10.130.0.59/23] from ovn-kubernetes logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:55 +0000 UTC Normal Pod report-span-q5l4g.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:56 +0000 UTC Normal Pod check-span-twqhx.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 824ms (824ms including waiting) kubelet logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:56 +0000 UTC Normal Pod check-span-twqhx.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:56 +0000 UTC Normal Pod check-span-twqhx.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:56 +0000 UTC Normal Pod report-span-q5l4g.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 889ms (889ms including waiting) kubelet logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:56 +0000 UTC Normal Pod report-span-q5l4g.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:17:56 +0000 UTC Normal Pod report-span-q5l4g.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:18:07 | examples-with-sampling | 2024-09-26 07:18:06 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:18:07 | examples-with-sampling | Deleting namespace: kuttl-test-dynamic-camel === CONT kuttl/harness/examples-with-cassandra logger.go:42: 07:21:32 | examples-with-cassandra | Creating namespace: kuttl-test-noted-fawn logger.go:42: 07:21:33 | examples-with-cassandra/0-install | starting test step 0-install logger.go:42: 07:21:33 | examples-with-cassandra/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 07:21:33 | examples-with-cassandra/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 07:21:33 | examples-with-cassandra/0-install | >>>> Creating namespace kuttl-test-noted-fawn logger.go:42: 07:21:33 | examples-with-cassandra/0-install | kubectl create namespace kuttl-test-noted-fawn 2>&1 | grep -v "already exists" || true logger.go:42: 07:21:33 | examples-with-cassandra/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-noted-fawn 2>&1 | grep -v "already exists" || true logger.go:42: 07:21:33 | examples-with-cassandra/0-install | service/cassandra created logger.go:42: 07:21:33 | examples-with-cassandra/0-install | statefulset.apps/cassandra created logger.go:42: 07:21:33 | examples-with-cassandra/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 07:21:37 | examples-with-cassandra/0-install | test step completed 0-install logger.go:42: 07:21:37 | examples-with-cassandra/1-install | starting test step 1-install logger.go:42: 07:21:37 | examples-with-cassandra/1-install | Jaeger:kuttl-test-noted-fawn/with-cassandra created logger.go:42: 07:22:06 | examples-with-cassandra/1-install | test step completed 1-install logger.go:42: 07:22:06 | examples-with-cassandra/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:22:06 | examples-with-cassandra/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-cassandra /dev/null] logger.go:42: 07:22:07 | examples-with-cassandra/2-smoke-test | Warning: resource jaegers/with-cassandra is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:22:51 | examples-with-cassandra/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:22:52 | examples-with-cassandra/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:22:52 | examples-with-cassandra/2-smoke-test | job.batch/report-span created logger.go:42: 07:22:52 | examples-with-cassandra/2-smoke-test | job.batch/check-span created logger.go:42: 07:23:05 | examples-with-cassandra/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:23:06 | examples-with-cassandra | examples-with-cassandra events from ns kuttl-test-noted-fawn: logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:21:34 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-noted-fawn/cassandra-0 to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:21:34 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.128.0.49/23] from ovn-kubernetes logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:21:34 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:21:34 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:21:34 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:21:34 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:21:35 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-noted-fawn/cassandra-1 to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:21:35 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:21:36 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.129.0.74/23] from ovn-kubernetes logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:21:36 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:21:36 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:21:36 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:21:38 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-q7jtj Binding Scheduled Successfully assigned kuttl-test-noted-fawn/with-cassandra-cassandra-schema-job-q7jtj to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:21:38 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-q7jtj AddedInterface Add eth0 [10.129.0.75/23] from ovn-kubernetes logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:21:38 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-q7jtj.spec.containers{with-cassandra-cassandra-schema-job} Pulling Pulling image "jaegertracing/jaeger-cassandra-schema:1.57.0" kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:21:38 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job SuccessfulCreate Created pod: with-cassandra-cassandra-schema-job-q7jtj job-controller logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:21:46 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-q7jtj.spec.containers{with-cassandra-cassandra-schema-job} Pulled Successfully pulled image "jaegertracing/jaeger-cassandra-schema:1.57.0" in 7.705s (7.705s including waiting) kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:21:46 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-q7jtj.spec.containers{with-cassandra-cassandra-schema-job} Created Created container with-cassandra-cassandra-schema-job kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:21:46 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-q7jtj.spec.containers{with-cassandra-cassandra-schema-job} Started Started container with-cassandra-cassandra-schema-job kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:21:52 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job Completed Job completed job-controller logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:21:53 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-789bc889c8 to 1 deployment-controller logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:03 +0000 UTC Normal Pod with-cassandra-789bc889c8-fv66b Binding Scheduled Successfully assigned kuttl-test-noted-fawn/with-cassandra-789bc889c8-fv66b to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:03 +0000 UTC Normal Pod with-cassandra-789bc889c8-fv66b AddedInterface Add eth0 [10.128.0.50/23] from ovn-kubernetes logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:03 +0000 UTC Normal Pod with-cassandra-789bc889c8-fv66b.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:03 +0000 UTC Normal Pod with-cassandra-789bc889c8-fv66b.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:03 +0000 UTC Normal Pod with-cassandra-789bc889c8-fv66b.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:03 +0000 UTC Normal Pod with-cassandra-789bc889c8-fv66b.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:03 +0000 UTC Normal ReplicaSet.apps with-cassandra-789bc889c8 SuccessfulCreate Created pod: with-cassandra-789bc889c8-fv66b replicaset-controller logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:04 +0000 UTC Normal Pod with-cassandra-789bc889c8-fv66b.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:04 +0000 UTC Normal Pod with-cassandra-789bc889c8-fv66b.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:08 +0000 UTC Normal Pod with-cassandra-789bc889c8-fv66b.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:08 +0000 UTC Normal Pod with-cassandra-789bc889c8-fv66b.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:08 +0000 UTC Normal ReplicaSet.apps with-cassandra-789bc889c8 SuccessfulDelete Deleted pod: with-cassandra-789bc889c8-fv66b replicaset-controller logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:08 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled down replica set with-cassandra-789bc889c8 to 0 from 1 deployment-controller logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:09 +0000 UTC Normal Pod with-cassandra-668bc9b89d-k6ssv Binding Scheduled Successfully assigned kuttl-test-noted-fawn/with-cassandra-668bc9b89d-k6ssv to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:09 +0000 UTC Normal ReplicaSet.apps with-cassandra-668bc9b89d SuccessfulCreate Created pod: with-cassandra-668bc9b89d-k6ssv replicaset-controller logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:09 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-668bc9b89d to 1 deployment-controller logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:10 +0000 UTC Normal Pod with-cassandra-668bc9b89d-k6ssv AddedInterface Add eth0 [10.129.0.76/23] from ovn-kubernetes logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:10 +0000 UTC Normal Pod with-cassandra-668bc9b89d-k6ssv.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:10 +0000 UTC Normal Pod with-cassandra-668bc9b89d-k6ssv.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:10 +0000 UTC Normal Pod with-cassandra-668bc9b89d-k6ssv.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:10 +0000 UTC Normal Pod with-cassandra-668bc9b89d-k6ssv.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:10 +0000 UTC Normal Pod with-cassandra-668bc9b89d-k6ssv.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:10 +0000 UTC Normal Pod with-cassandra-668bc9b89d-k6ssv.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:11 +0000 UTC Warning Pod with-cassandra-668bc9b89d-k6ssv.spec.containers{jaeger} BackOff Back-off restarting failed container jaeger in pod with-cassandra-668bc9b89d-k6ssv_kuttl-test-noted-fawn(0710de54-776f-429e-aef7-9589f5b96d91) kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:52 +0000 UTC Normal Pod check-span-b56cx Binding Scheduled Successfully assigned kuttl-test-noted-fawn/check-span-b56cx to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:52 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-b56cx job-controller logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:52 +0000 UTC Normal Pod report-span-ljq8j Binding Scheduled Successfully assigned kuttl-test-noted-fawn/report-span-ljq8j to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:52 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-ljq8j job-controller logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:53 +0000 UTC Normal Pod check-span-b56cx AddedInterface Add eth0 [10.128.0.51/23] from ovn-kubernetes logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:53 +0000 UTC Normal Pod check-span-b56cx.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:53 +0000 UTC Normal Pod report-span-ljq8j AddedInterface Add eth0 [10.130.0.62/23] from ovn-kubernetes logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:53 +0000 UTC Normal Pod report-span-ljq8j.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:54 +0000 UTC Normal Pod check-span-b56cx.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 954ms (954ms including waiting) kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:54 +0000 UTC Normal Pod check-span-b56cx.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:54 +0000 UTC Normal Pod check-span-b56cx.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:54 +0000 UTC Normal Pod report-span-ljq8j.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 997ms (997ms including waiting) kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:54 +0000 UTC Normal Pod report-span-ljq8j.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:22:54 +0000 UTC Normal Pod report-span-ljq8j.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:23:06 | examples-with-cassandra | 2024-09-26 07:23:05 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:23:06 | examples-with-cassandra | Deleting namespace: kuttl-test-noted-fawn === CONT kuttl/harness/examples-with-badger logger.go:42: 07:23:18 | examples-with-badger | Creating namespace: kuttl-test-amused-minnow logger.go:42: 07:23:18 | examples-with-badger/0-install | starting test step 0-install logger.go:42: 07:23:18 | examples-with-badger/0-install | Jaeger:kuttl-test-amused-minnow/with-badger created logger.go:42: 07:23:20 | examples-with-badger/0-install | test step completed 0-install logger.go:42: 07:23:20 | examples-with-badger/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:23:20 | examples-with-badger/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-badger /dev/null] logger.go:42: 07:23:22 | examples-with-badger/1-smoke-test | Warning: resource jaegers/with-badger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:23:29 | examples-with-badger/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:23:30 | examples-with-badger/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:23:30 | examples-with-badger/1-smoke-test | job.batch/report-span created logger.go:42: 07:23:30 | examples-with-badger/1-smoke-test | job.batch/check-span created logger.go:42: 07:23:43 | examples-with-badger/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:23:44 | examples-with-badger | examples-with-badger events from ns kuttl-test-amused-minnow: logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:19 +0000 UTC Normal Pod with-badger-7fd858f6b9-l494l Binding Scheduled Successfully assigned kuttl-test-amused-minnow/with-badger-7fd858f6b9-l494l to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:19 +0000 UTC Normal Pod with-badger-7fd858f6b9-l494l AddedInterface Add eth0 [10.130.0.63/23] from ovn-kubernetes logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:19 +0000 UTC Normal Pod with-badger-7fd858f6b9-l494l.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:19 +0000 UTC Normal Pod with-badger-7fd858f6b9-l494l.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:19 +0000 UTC Normal Pod with-badger-7fd858f6b9-l494l.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:19 +0000 UTC Normal Pod with-badger-7fd858f6b9-l494l.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:19 +0000 UTC Normal Pod with-badger-7fd858f6b9-l494l.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:19 +0000 UTC Normal Pod with-badger-7fd858f6b9-l494l.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:19 +0000 UTC Normal ReplicaSet.apps with-badger-7fd858f6b9 SuccessfulCreate Created pod: with-badger-7fd858f6b9-l494l replicaset-controller logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:19 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-7fd858f6b9 to 1 deployment-controller logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:23 +0000 UTC Normal Pod with-badger-7fd858f6b9-l494l.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:23 +0000 UTC Normal Pod with-badger-7fd858f6b9-l494l.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:23 +0000 UTC Normal ReplicaSet.apps with-badger-7fd858f6b9 SuccessfulDelete Deleted pod: with-badger-7fd858f6b9-l494l replicaset-controller logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:23 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled down replica set with-badger-7fd858f6b9 to 0 from 1 deployment-controller logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:24 +0000 UTC Normal Pod with-badger-84d898ffc8-nbmdj Binding Scheduled Successfully assigned kuttl-test-amused-minnow/with-badger-84d898ffc8-nbmdj to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:24 +0000 UTC Normal Pod with-badger-84d898ffc8-nbmdj AddedInterface Add eth0 [10.129.0.77/23] from ovn-kubernetes logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:24 +0000 UTC Normal Pod with-badger-84d898ffc8-nbmdj.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:24 +0000 UTC Normal ReplicaSet.apps with-badger-84d898ffc8 SuccessfulCreate Created pod: with-badger-84d898ffc8-nbmdj replicaset-controller logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:24 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-84d898ffc8 to 1 deployment-controller logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:25 +0000 UTC Normal Pod with-badger-84d898ffc8-nbmdj.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:25 +0000 UTC Normal Pod with-badger-84d898ffc8-nbmdj.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:25 +0000 UTC Normal Pod with-badger-84d898ffc8-nbmdj.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:25 +0000 UTC Normal Pod with-badger-84d898ffc8-nbmdj.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:25 +0000 UTC Normal Pod with-badger-84d898ffc8-nbmdj.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:30 +0000 UTC Normal Pod check-span-2pml6 Binding Scheduled Successfully assigned kuttl-test-amused-minnow/check-span-2pml6 to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:30 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-2pml6 job-controller logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:30 +0000 UTC Normal Pod report-span-475z9 Binding Scheduled Successfully assigned kuttl-test-amused-minnow/report-span-475z9 to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:30 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-475z9 job-controller logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:31 +0000 UTC Normal Pod check-span-2pml6 AddedInterface Add eth0 [10.128.0.53/23] from ovn-kubernetes logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:31 +0000 UTC Normal Pod check-span-2pml6.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:31 +0000 UTC Normal Pod report-span-475z9 AddedInterface Add eth0 [10.128.0.52/23] from ovn-kubernetes logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:31 +0000 UTC Normal Pod report-span-475z9.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:32 +0000 UTC Normal Pod check-span-2pml6.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 871ms (871ms including waiting) kubelet logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:32 +0000 UTC Normal Pod check-span-2pml6.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:32 +0000 UTC Normal Pod check-span-2pml6.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:32 +0000 UTC Normal Pod report-span-475z9.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 818ms (818ms including waiting) kubelet logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:32 +0000 UTC Normal Pod report-span-475z9.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:32 +0000 UTC Normal Pod report-span-475z9.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:23:44 | examples-with-badger | 2024-09-26 07:23:43 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:23:44 | examples-with-badger | Deleting namespace: kuttl-test-amused-minnow === CONT kuttl/harness/examples-simplest logger.go:42: 07:23:55 | examples-simplest | Creating namespace: kuttl-test-perfect-possum logger.go:42: 07:23:55 | examples-simplest/0-install | starting test step 0-install logger.go:42: 07:23:56 | examples-simplest/0-install | Jaeger:kuttl-test-perfect-possum/simplest created logger.go:42: 07:23:59 | examples-simplest/0-install | test step completed 0-install logger.go:42: 07:23:59 | examples-simplest/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:23:59 | examples-simplest/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 07:24:01 | examples-simplest/1-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:24:08 | examples-simplest/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:24:09 | examples-simplest/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:24:09 | examples-simplest/1-smoke-test | job.batch/report-span created logger.go:42: 07:24:09 | examples-simplest/1-smoke-test | job.batch/check-span created logger.go:42: 07:24:22 | examples-simplest/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:24:22 | examples-simplest | examples-simplest events from ns kuttl-test-perfect-possum: logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:23:56 +0000 UTC Normal Pod simplest-85864f5444-jbtnb Binding Scheduled Successfully assigned kuttl-test-perfect-possum/simplest-85864f5444-jbtnb to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:23:56 +0000 UTC Normal ReplicaSet.apps simplest-85864f5444 SuccessfulCreate Created pod: simplest-85864f5444-jbtnb replicaset-controller logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:23:56 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-85864f5444 to 1 deployment-controller logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:23:57 +0000 UTC Warning Pod simplest-85864f5444-jbtnb FailedMount MountVolume.SetUp failed for volume "simplest-collector-tls-config-volume" : secret "simplest-collector-headless-tls" not found kubelet logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:23:57 +0000 UTC Normal Pod simplest-85864f5444-jbtnb AddedInterface Add eth0 [10.129.0.78/23] from ovn-kubernetes logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:23:57 +0000 UTC Normal Pod simplest-85864f5444-jbtnb.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:23:57 +0000 UTC Normal Pod simplest-85864f5444-jbtnb.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:23:57 +0000 UTC Normal Pod simplest-85864f5444-jbtnb.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:23:57 +0000 UTC Normal Pod simplest-85864f5444-jbtnb.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:23:58 +0000 UTC Normal Pod simplest-85864f5444-jbtnb.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:23:58 +0000 UTC Normal Pod simplest-85864f5444-jbtnb.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:24:02 +0000 UTC Normal Pod simplest-7d659577bb-r7nws Binding Scheduled Successfully assigned kuttl-test-perfect-possum/simplest-7d659577bb-r7nws to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:24:02 +0000 UTC Normal ReplicaSet.apps simplest-7d659577bb SuccessfulCreate Created pod: simplest-7d659577bb-r7nws replicaset-controller logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:24:02 +0000 UTC Normal Pod simplest-85864f5444-jbtnb.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:24:02 +0000 UTC Normal Pod simplest-85864f5444-jbtnb.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:24:02 +0000 UTC Normal ReplicaSet.apps simplest-85864f5444 SuccessfulDelete Deleted pod: simplest-85864f5444-jbtnb replicaset-controller logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:24:02 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-85864f5444 to 0 from 1 deployment-controller logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:24:02 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-7d659577bb to 1 deployment-controller logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:24:03 +0000 UTC Normal Pod simplest-7d659577bb-r7nws AddedInterface Add eth0 [10.129.0.79/23] from ovn-kubernetes logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:24:03 +0000 UTC Normal Pod simplest-7d659577bb-r7nws.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:24:03 +0000 UTC Normal Pod simplest-7d659577bb-r7nws.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:24:03 +0000 UTC Normal Pod simplest-7d659577bb-r7nws.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:24:03 +0000 UTC Normal Pod simplest-7d659577bb-r7nws.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:24:03 +0000 UTC Normal Pod simplest-7d659577bb-r7nws.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:24:03 +0000 UTC Normal Pod simplest-7d659577bb-r7nws.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:24:09 +0000 UTC Normal Pod check-span-frm75 Binding Scheduled Successfully assigned kuttl-test-perfect-possum/check-span-frm75 to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:24:09 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-frm75 job-controller logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:24:09 +0000 UTC Normal Pod report-span-kqkws Binding Scheduled Successfully assigned kuttl-test-perfect-possum/report-span-kqkws to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:24:09 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-kqkws job-controller logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:24:10 +0000 UTC Normal Pod check-span-frm75 AddedInterface Add eth0 [10.130.0.64/23] from ovn-kubernetes logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:24:10 +0000 UTC Normal Pod check-span-frm75.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:24:10 +0000 UTC Normal Pod report-span-kqkws AddedInterface Add eth0 [10.128.0.54/23] from ovn-kubernetes logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:24:10 +0000 UTC Normal Pod report-span-kqkws.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:24:10 +0000 UTC Normal Pod report-span-kqkws.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 816ms (816ms including waiting) kubelet logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:24:10 +0000 UTC Normal Pod report-span-kqkws.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:24:11 +0000 UTC Normal Pod check-span-frm75.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 845ms (845ms including waiting) kubelet logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:24:11 +0000 UTC Normal Pod check-span-frm75.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:24:11 +0000 UTC Normal Pod check-span-frm75.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:24:11 +0000 UTC Normal Pod report-span-kqkws.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:24:22 | examples-simplest | 2024-09-26 07:24:22 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:24:22 | examples-simplest | Deleting namespace: kuttl-test-perfect-possum === CONT kuttl/harness/examples-simple-prod-with-volumes logger.go:42: 07:24:34 | examples-simple-prod-with-volumes | Ignoring 03-check-volume.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:24:34 | examples-simple-prod-with-volumes | Creating namespace: kuttl-test-many-puma logger.go:42: 07:24:34 | examples-simple-prod-with-volumes/1-install | starting test step 1-install logger.go:42: 07:24:35 | examples-simple-prod-with-volumes/1-install | Jaeger:kuttl-test-many-puma/simple-prod created logger.go:42: 07:25:09 | examples-simple-prod-with-volumes/1-install | test step completed 1-install logger.go:42: 07:25:09 | examples-simple-prod-with-volumes/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:25:09 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:25:11 | examples-simple-prod-with-volumes/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:25:18 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:25:19 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:25:19 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/report-span created logger.go:42: 07:25:19 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/check-span created logger.go:42: 07:25:32 | examples-simple-prod-with-volumes/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:25:32 | examples-simple-prod-with-volumes/3-check-volume | starting test step 3-check-volume logger.go:42: 07:25:32 | examples-simple-prod-with-volumes/3-check-volume | running command: [sh -c kubectl exec $(kubectl get pods -n $NAMESPACE -l app=jaeger -l app.kubernetes.io/component=collector -o yaml | /tmp/jaeger-tests/bin/yq e '.items[0].metadata.name') -n $NAMESPACE -- ls /usr/share/elasticsearch/data] logger.go:42: 07:25:34 | examples-simple-prod-with-volumes/3-check-volume | test step completed 3-check-volume logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | examples-simple-prod-with-volumes events from ns kuttl-test-many-puma: logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:24:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmanypumasimpleprod-1-9bf7fb687-4zk58 Binding Scheduled Successfully assigned kuttl-test-many-puma/elasticsearch-cdm-kuttltestmanypumasimpleprod-1-9bf7fb687-4zk58 to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:24:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmanypumasimpleprod-1-9bf7fb687-4zk58 AddedInterface Add eth0 [10.128.0.55/23] from ovn-kubernetes logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:24:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmanypumasimpleprod-1-9bf7fb687-4zk58.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:bc97b8e13087050dfb1a75b02c5b780cbb8fb12a849a655169d072ba8bbf42b4" already present on machine kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:24:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmanypumasimpleprod-1-9bf7fb687-4zk58.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:24:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmanypumasimpleprod-1-9bf7fb687-4zk58.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:24:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmanypumasimpleprod-1-9bf7fb687-4zk58.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:1d1d7d47b616995d18692f3a2d6232d7e1a0f41bc1503d87100f31939e080a78" already present on machine kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:24:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmanypumasimpleprod-1-9bf7fb687-4zk58.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:24:39 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestmanypumasimpleprod-1-9bf7fb687 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestmanypumasimpleprod-1-9bf7fb687-4zk58 replicaset-controller logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:24:39 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestmanypumasimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestmanypumasimpleprod-1-9bf7fb687 to 1 deployment-controller logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:24:40 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmanypumasimpleprod-1-9bf7fb687-4zk58.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:24:49 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmanypumasimpleprod-1-9bf7fb687-4zk58.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:24:54 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmanypumasimpleprod-1-9bf7fb687-4zk58.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:06 +0000 UTC Normal Pod simple-prod-collector-65f55d446f-q77cg Binding Scheduled Successfully assigned kuttl-test-many-puma/simple-prod-collector-65f55d446f-q77cg to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:06 +0000 UTC Normal Pod simple-prod-collector-65f55d446f-q77cg AddedInterface Add eth0 [10.129.0.80/23] from ovn-kubernetes logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:06 +0000 UTC Normal Pod simple-prod-collector-65f55d446f-q77cg.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:06 +0000 UTC Normal Pod simple-prod-collector-65f55d446f-q77cg.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:06 +0000 UTC Normal Pod simple-prod-collector-65f55d446f-q77cg.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:06 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-65f55d446f SuccessfulCreate Created pod: simple-prod-collector-65f55d446f-q77cg replicaset-controller logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:06 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-65f55d446f to 1 deployment-controller logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:06 +0000 UTC Normal Pod simple-prod-query-74dc5dd9d9-cl9bs Binding Scheduled Successfully assigned kuttl-test-many-puma/simple-prod-query-74dc5dd9d9-cl9bs to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:06 +0000 UTC Warning Pod simple-prod-query-74dc5dd9d9-cl9bs FailedMount MountVolume.SetUp failed for volume "simple-prod-ui-oauth-proxy-tls" : secret "simple-prod-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:06 +0000 UTC Normal ReplicaSet.apps simple-prod-query-74dc5dd9d9 SuccessfulCreate Created pod: simple-prod-query-74dc5dd9d9-cl9bs replicaset-controller logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:06 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-74dc5dd9d9 to 1 deployment-controller logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:07 +0000 UTC Normal Pod simple-prod-query-74dc5dd9d9-cl9bs AddedInterface Add eth0 [10.130.0.65/23] from ovn-kubernetes logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:07 +0000 UTC Normal Pod simple-prod-query-74dc5dd9d9-cl9bs.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:07 +0000 UTC Normal Pod simple-prod-query-74dc5dd9d9-cl9bs.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:07 +0000 UTC Normal Pod simple-prod-query-74dc5dd9d9-cl9bs.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:07 +0000 UTC Normal Pod simple-prod-query-74dc5dd9d9-cl9bs.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:07 +0000 UTC Normal Pod simple-prod-query-74dc5dd9d9-cl9bs.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:07 +0000 UTC Normal Pod simple-prod-query-74dc5dd9d9-cl9bs.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:07 +0000 UTC Normal Pod simple-prod-query-74dc5dd9d9-cl9bs.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:07 +0000 UTC Normal Pod simple-prod-query-74dc5dd9d9-cl9bs.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:07 +0000 UTC Normal Pod simple-prod-query-74dc5dd9d9-cl9bs.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:12 +0000 UTC Normal Pod simple-prod-query-74dc5dd9d9-cl9bs.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:12 +0000 UTC Normal Pod simple-prod-query-74dc5dd9d9-cl9bs.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:12 +0000 UTC Normal Pod simple-prod-query-74dc5dd9d9-cl9bs.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:12 +0000 UTC Normal ReplicaSet.apps simple-prod-query-74dc5dd9d9 SuccessfulDelete Deleted pod: simple-prod-query-74dc5dd9d9-cl9bs replicaset-controller logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:12 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-74dc5dd9d9 to 0 from 1 deployment-controller logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:13 +0000 UTC Normal Pod simple-prod-query-76c8c645c6-hzgf4 Binding Scheduled Successfully assigned kuttl-test-many-puma/simple-prod-query-76c8c645c6-hzgf4 to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:13 +0000 UTC Normal Pod simple-prod-query-76c8c645c6-hzgf4 AddedInterface Add eth0 [10.130.0.66/23] from ovn-kubernetes logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:13 +0000 UTC Normal Pod simple-prod-query-76c8c645c6-hzgf4.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:13 +0000 UTC Normal Pod simple-prod-query-76c8c645c6-hzgf4.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:13 +0000 UTC Normal Pod simple-prod-query-76c8c645c6-hzgf4.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:13 +0000 UTC Normal Pod simple-prod-query-76c8c645c6-hzgf4.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:13 +0000 UTC Normal Pod simple-prod-query-76c8c645c6-hzgf4.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:13 +0000 UTC Normal Pod simple-prod-query-76c8c645c6-hzgf4.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:13 +0000 UTC Normal Pod simple-prod-query-76c8c645c6-hzgf4.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:13 +0000 UTC Normal Pod simple-prod-query-76c8c645c6-hzgf4.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:13 +0000 UTC Normal Pod simple-prod-query-76c8c645c6-hzgf4.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:13 +0000 UTC Normal ReplicaSet.apps simple-prod-query-76c8c645c6 SuccessfulCreate Created pod: simple-prod-query-76c8c645c6-hzgf4 replicaset-controller logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:13 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-76c8c645c6 to 1 deployment-controller logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:19 +0000 UTC Normal Pod check-span-xsn2z Binding Scheduled Successfully assigned kuttl-test-many-puma/check-span-xsn2z to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:19 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-xsn2z job-controller logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:19 +0000 UTC Normal Pod report-span-7jwq8 Binding Scheduled Successfully assigned kuttl-test-many-puma/report-span-7jwq8 to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:19 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-7jwq8 job-controller logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:20 +0000 UTC Normal Pod check-span-xsn2z AddedInterface Add eth0 [10.129.0.82/23] from ovn-kubernetes logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:20 +0000 UTC Normal Pod check-span-xsn2z.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:20 +0000 UTC Normal Pod report-span-7jwq8 AddedInterface Add eth0 [10.129.0.81/23] from ovn-kubernetes logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:20 +0000 UTC Normal Pod report-span-7jwq8.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:21 +0000 UTC Normal Pod check-span-xsn2z.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 811ms (811ms including waiting) kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:21 +0000 UTC Normal Pod check-span-xsn2z.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:21 +0000 UTC Normal Pod check-span-xsn2z.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:21 +0000 UTC Normal Pod report-span-7jwq8.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 934ms (934ms including waiting) kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:21 +0000 UTC Normal Pod report-span-7jwq8.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:21 +0000 UTC Normal Pod report-span-7jwq8.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:21 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:21 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:21 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | 2024-09-26 07:25:32 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:25:34 | examples-simple-prod-with-volumes | Deleting namespace: kuttl-test-many-puma === CONT kuttl/harness/examples-business-application-injected-sidecar logger.go:42: 07:25:41 | examples-business-application-injected-sidecar | Creating namespace: kuttl-test-welcomed-lion logger.go:42: 07:25:41 | examples-business-application-injected-sidecar/0-install | starting test step 0-install logger.go:42: 07:25:41 | examples-business-application-injected-sidecar/0-install | Deployment:kuttl-test-welcomed-lion/myapp created logger.go:42: 07:25:41 | examples-business-application-injected-sidecar/0-install | test step completed 0-install logger.go:42: 07:25:41 | examples-business-application-injected-sidecar/1-install | starting test step 1-install logger.go:42: 07:25:41 | examples-business-application-injected-sidecar/1-install | Jaeger:kuttl-test-welcomed-lion/simplest created logger.go:42: 07:25:46 | examples-business-application-injected-sidecar/1-install | test step completed 1-install logger.go:42: 07:25:46 | examples-business-application-injected-sidecar/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:25:46 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 07:25:48 | examples-business-application-injected-sidecar/2-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:25:55 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:25:55 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:25:56 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/report-span created logger.go:42: 07:25:56 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/check-span created logger.go:42: 07:26:09 | examples-business-application-injected-sidecar/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | examples-business-application-injected-sidecar events from ns kuttl-test-welcomed-lion: logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:41 +0000 UTC Normal Pod myapp-679f79d5f8-9js45 Binding Scheduled Successfully assigned kuttl-test-welcomed-lion/myapp-679f79d5f8-9js45 to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:41 +0000 UTC Normal Pod myapp-679f79d5f8-9js45 AddedInterface Add eth0 [10.128.0.56/23] from ovn-kubernetes logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:41 +0000 UTC Normal ReplicaSet.apps myapp-679f79d5f8 SuccessfulCreate Created pod: myapp-679f79d5f8-9js45 replicaset-controller logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:41 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-679f79d5f8 to 1 deployment-controller logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:42 +0000 UTC Normal Pod myapp-679f79d5f8-9js45.spec.containers{myapp} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:42 +0000 UTC Normal Pod myapp-85cdf4fd8c-c7kb2 Binding Scheduled Successfully assigned kuttl-test-welcomed-lion/myapp-85cdf4fd8c-c7kb2 to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:42 +0000 UTC Warning Pod myapp-85cdf4fd8c-c7kb2 FailedMount MountVolume.SetUp failed for volume "simplest-service-ca" : configmap "simplest-service-ca" not found kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:42 +0000 UTC Warning Pod myapp-85cdf4fd8c-c7kb2 FailedMount MountVolume.SetUp failed for volume "simplest-trusted-ca" : configmap "simplest-trusted-ca" not found kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:42 +0000 UTC Normal ReplicaSet.apps myapp-85cdf4fd8c SuccessfulCreate Created pod: myapp-85cdf4fd8c-c7kb2 replicaset-controller logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:42 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-85cdf4fd8c to 1 deployment-controller logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:43 +0000 UTC Normal Pod myapp-85cdf4fd8c-c7kb2 AddedInterface Add eth0 [10.128.0.57/23] from ovn-kubernetes logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:43 +0000 UTC Normal Pod simplest-b56df6945-hvc8c Binding Scheduled Successfully assigned kuttl-test-welcomed-lion/simplest-b56df6945-hvc8c to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:43 +0000 UTC Normal Pod simplest-b56df6945-hvc8c AddedInterface Add eth0 [10.130.0.67/23] from ovn-kubernetes logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:43 +0000 UTC Normal Pod simplest-b56df6945-hvc8c.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:43 +0000 UTC Normal Pod simplest-b56df6945-hvc8c.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:43 +0000 UTC Normal Pod simplest-b56df6945-hvc8c.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:43 +0000 UTC Normal Pod simplest-b56df6945-hvc8c.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:43 +0000 UTC Normal ReplicaSet.apps simplest-b56df6945 SuccessfulCreate Created pod: simplest-b56df6945-hvc8c replicaset-controller logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:43 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-b56df6945 to 1 deployment-controller logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:44 +0000 UTC Normal Pod myapp-85cdf4fd8c-c7kb2.spec.containers{myapp} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:44 +0000 UTC Normal Pod simplest-b56df6945-hvc8c.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:44 +0000 UTC Normal Pod simplest-b56df6945-hvc8c.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:47 +0000 UTC Normal Pod myapp-679f79d5f8-9js45.spec.containers{myapp} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 5.316s (5.316s including waiting) kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:47 +0000 UTC Normal Pod myapp-679f79d5f8-9js45.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:47 +0000 UTC Normal Pod myapp-679f79d5f8-9js45.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:47 +0000 UTC Normal Pod myapp-85cdf4fd8c-c7kb2.spec.containers{myapp} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 3.531s (3.531s including waiting) kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:47 +0000 UTC Normal Pod myapp-85cdf4fd8c-c7kb2.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:47 +0000 UTC Normal Pod myapp-85cdf4fd8c-c7kb2.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:47 +0000 UTC Normal Pod myapp-85cdf4fd8c-c7kb2.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:47 +0000 UTC Normal Pod myapp-85cdf4fd8c-c7kb2.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:47 +0000 UTC Normal Pod myapp-85cdf4fd8c-c7kb2.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:48 +0000 UTC Normal Pod simplest-b56df6945-hvc8c.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:48 +0000 UTC Normal Pod simplest-b56df6945-hvc8c.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:48 +0000 UTC Normal ReplicaSet.apps simplest-b56df6945 SuccessfulDelete Deleted pod: simplest-b56df6945-hvc8c replicaset-controller logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:48 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-b56df6945 to 0 from 1 deployment-controller logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:49 +0000 UTC Normal ReplicaSet.apps myapp-679f79d5f8 SuccessfulDelete Deleted pod: myapp-679f79d5f8-9js45 replicaset-controller logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:49 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled down replica set myapp-679f79d5f8 to 0 from 1 deployment-controller logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:49 +0000 UTC Normal Pod simplest-5db57f6cc-47wls Binding Scheduled Successfully assigned kuttl-test-welcomed-lion/simplest-5db57f6cc-47wls to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:49 +0000 UTC Normal ReplicaSet.apps simplest-5db57f6cc SuccessfulCreate Created pod: simplest-5db57f6cc-47wls replicaset-controller logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:49 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-5db57f6cc to 1 deployment-controller logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:50 +0000 UTC Normal Pod myapp-679f79d5f8-9js45.spec.containers{myapp} Killing Stopping container myapp kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:50 +0000 UTC Normal Pod simplest-5db57f6cc-47wls AddedInterface Add eth0 [10.129.0.83/23] from ovn-kubernetes logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:50 +0000 UTC Normal Pod simplest-5db57f6cc-47wls.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:50 +0000 UTC Normal Pod simplest-5db57f6cc-47wls.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:50 +0000 UTC Normal Pod simplest-5db57f6cc-47wls.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:50 +0000 UTC Normal Pod simplest-5db57f6cc-47wls.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:50 +0000 UTC Normal Pod simplest-5db57f6cc-47wls.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:50 +0000 UTC Normal Pod simplest-5db57f6cc-47wls.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:54 +0000 UTC Warning Pod myapp-85cdf4fd8c-c7kb2.spec.containers{myapp} Unhealthy Liveness probe failed: Get "http://10.128.0.57:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:56 +0000 UTC Normal Pod check-span-g6mcc Binding Scheduled Successfully assigned kuttl-test-welcomed-lion/check-span-g6mcc to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:56 +0000 UTC Normal Pod check-span-g6mcc AddedInterface Add eth0 [10.130.0.69/23] from ovn-kubernetes logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:56 +0000 UTC Normal Pod check-span-g6mcc.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:56 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-g6mcc job-controller logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:56 +0000 UTC Normal Pod report-span-6bfr8 Binding Scheduled Successfully assigned kuttl-test-welcomed-lion/report-span-6bfr8 to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:56 +0000 UTC Normal Pod report-span-6bfr8 AddedInterface Add eth0 [10.130.0.68/23] from ovn-kubernetes logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:56 +0000 UTC Normal Pod report-span-6bfr8.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:56 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-6bfr8 job-controller logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:57 +0000 UTC Normal Pod check-span-g6mcc.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 840ms (840ms including waiting) kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:57 +0000 UTC Normal Pod check-span-g6mcc.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:57 +0000 UTC Normal Pod check-span-g6mcc.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:57 +0000 UTC Normal Pod report-span-6bfr8.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 845ms (845ms including waiting) kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:57 +0000 UTC Normal Pod report-span-6bfr8.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:25:57 +0000 UTC Normal Pod report-span-6bfr8.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | 2024-09-26 07:26:08 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:26:09 | examples-business-application-injected-sidecar | Deleting namespace: kuttl-test-welcomed-lion === CONT kuttl/harness/examples-service-types logger.go:42: 07:26:21 | examples-service-types | Creating namespace: kuttl-test-prompt-titmouse logger.go:42: 07:26:21 | examples-service-types/0-install | starting test step 0-install logger.go:42: 07:26:22 | examples-service-types/0-install | Jaeger:kuttl-test-prompt-titmouse/service-types created logger.go:42: 07:26:25 | examples-service-types/0-install | test step completed 0-install logger.go:42: 07:26:25 | examples-service-types/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:26:25 | examples-service-types/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE service-types /dev/null] logger.go:42: 07:26:27 | examples-service-types/1-smoke-test | Warning: resource jaegers/service-types is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:26:34 | examples-service-types/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://service-types-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:26:34 | examples-service-types/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:26:35 | examples-service-types/1-smoke-test | job.batch/report-span created logger.go:42: 07:26:35 | examples-service-types/1-smoke-test | job.batch/check-span created logger.go:42: 07:26:49 | examples-service-types/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:26:49 | examples-service-types/2- | starting test step 2- logger.go:42: 07:26:50 | examples-service-types/2- | test step completed 2- logger.go:42: 07:26:50 | examples-service-types | examples-service-types events from ns kuttl-test-prompt-titmouse: logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:22 +0000 UTC Normal Pod service-types-7f55ff6c9b-zlmtk Binding Scheduled Successfully assigned kuttl-test-prompt-titmouse/service-types-7f55ff6c9b-zlmtk to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:22 +0000 UTC Normal ReplicaSet.apps service-types-7f55ff6c9b SuccessfulCreate Created pod: service-types-7f55ff6c9b-zlmtk replicaset-controller logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:22 +0000 UTC Normal Service service-types-collector EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:22 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-7f55ff6c9b to 1 deployment-controller logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:23 +0000 UTC Normal Pod service-types-7f55ff6c9b-zlmtk AddedInterface Add eth0 [10.130.0.70/23] from ovn-kubernetes logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:23 +0000 UTC Normal Pod service-types-7f55ff6c9b-zlmtk.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:23 +0000 UTC Normal Pod service-types-7f55ff6c9b-zlmtk.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:23 +0000 UTC Normal Pod service-types-7f55ff6c9b-zlmtk.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:23 +0000 UTC Normal Pod service-types-7f55ff6c9b-zlmtk.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:23 +0000 UTC Normal Pod service-types-7f55ff6c9b-zlmtk.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:23 +0000 UTC Normal Pod service-types-7f55ff6c9b-zlmtk.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:25 +0000 UTC Normal Service service-types-collector EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:25 +0000 UTC Normal Service service-types-query EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:27 +0000 UTC Normal Pod service-types-7f55ff6c9b-zlmtk.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:27 +0000 UTC Normal Pod service-types-7f55ff6c9b-zlmtk.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:27 +0000 UTC Normal ReplicaSet.apps service-types-7f55ff6c9b SuccessfulDelete Deleted pod: service-types-7f55ff6c9b-zlmtk replicaset-controller logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:27 +0000 UTC Normal Service service-types-query EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:27 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled down replica set service-types-7f55ff6c9b to 0 from 1 deployment-controller logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:29 +0000 UTC Normal Pod service-types-7bd75d4b44-jwrn7 Binding Scheduled Successfully assigned kuttl-test-prompt-titmouse/service-types-7bd75d4b44-jwrn7 to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:29 +0000 UTC Normal Pod service-types-7bd75d4b44-jwrn7 AddedInterface Add eth0 [10.128.0.58/23] from ovn-kubernetes logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:29 +0000 UTC Normal Pod service-types-7bd75d4b44-jwrn7.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:29 +0000 UTC Normal Pod service-types-7bd75d4b44-jwrn7.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:29 +0000 UTC Normal Pod service-types-7bd75d4b44-jwrn7.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:29 +0000 UTC Normal Pod service-types-7bd75d4b44-jwrn7.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:29 +0000 UTC Normal Pod service-types-7bd75d4b44-jwrn7.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:29 +0000 UTC Normal Pod service-types-7bd75d4b44-jwrn7.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:29 +0000 UTC Normal ReplicaSet.apps service-types-7bd75d4b44 SuccessfulCreate Created pod: service-types-7bd75d4b44-jwrn7 replicaset-controller logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:29 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-7bd75d4b44 to 1 deployment-controller logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:35 +0000 UTC Normal Pod check-span-g9rlr Binding Scheduled Successfully assigned kuttl-test-prompt-titmouse/check-span-g9rlr to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:35 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-g9rlr job-controller logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:35 +0000 UTC Normal Pod report-span-rshsv Binding Scheduled Successfully assigned kuttl-test-prompt-titmouse/report-span-rshsv to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:35 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-rshsv job-controller logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:36 +0000 UTC Normal Pod check-span-g9rlr AddedInterface Add eth0 [10.130.0.71/23] from ovn-kubernetes logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:36 +0000 UTC Normal Pod check-span-g9rlr.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:36 +0000 UTC Normal Pod check-span-g9rlr.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 891ms (891ms including waiting) kubelet logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:36 +0000 UTC Normal Pod report-span-rshsv AddedInterface Add eth0 [10.128.0.59/23] from ovn-kubernetes logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:36 +0000 UTC Normal Pod report-span-rshsv.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:37 +0000 UTC Normal Pod check-span-g9rlr.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:37 +0000 UTC Normal Pod check-span-g9rlr.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:37 +0000 UTC Normal Pod report-span-rshsv.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 861ms (861ms including waiting) kubelet logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:37 +0000 UTC Normal Pod report-span-rshsv.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:37 +0000 UTC Normal Pod report-span-rshsv.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:26:50 | examples-service-types | 2024-09-26 07:26:49 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:26:50 | examples-service-types | Deleting namespace: kuttl-test-prompt-titmouse === CONT kuttl/harness/examples-openshift-with-htpasswd logger.go:42: 07:27:16 | examples-openshift-with-htpasswd | Ignoring 00-install.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:27:16 | examples-openshift-with-htpasswd | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:27:16 | examples-openshift-with-htpasswd | Creating namespace: kuttl-test-sound-panda logger.go:42: 07:27:16 | examples-openshift-with-htpasswd/0-install | starting test step 0-install logger.go:42: 07:27:16 | examples-openshift-with-htpasswd/0-install | Secret:kuttl-test-sound-panda/htpasswd created logger.go:42: 07:27:16 | examples-openshift-with-htpasswd/0-install | test step completed 0-install logger.go:42: 07:27:16 | examples-openshift-with-htpasswd/1-install | starting test step 1-install logger.go:42: 07:27:16 | examples-openshift-with-htpasswd/1-install | Jaeger:kuttl-test-sound-panda/with-htpasswd created logger.go:42: 07:27:20 | examples-openshift-with-htpasswd/1-install | test step completed 1-install logger.go:42: 07:27:20 | examples-openshift-with-htpasswd/2-check-unsecured | starting test step 2-check-unsecured logger.go:42: 07:27:20 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [./ensure-ingress-host.sh] logger.go:42: 07:27:20 | examples-openshift-with-htpasswd/2-check-unsecured | Checking the Ingress host value was populated logger.go:42: 07:27:20 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 0 logger.go:42: 07:27:20 | examples-openshift-with-htpasswd/2-check-unsecured | Hostname is with-htpasswd-kuttl-test-sound-panda.apps.rosa.ci-rosa-h-ezi9.l97m.s3.devshift.org logger.go:42: 07:27:20 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 07:27:20 | examples-openshift-with-htpasswd/2-check-unsecured | Checking an expected HTTP response logger.go:42: 07:27:20 | examples-openshift-with-htpasswd/2-check-unsecured | Running in OpenShift logger.go:42: 07:27:20 | examples-openshift-with-htpasswd/2-check-unsecured | Not using any secret logger.go:42: 07:27:20 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 1/30 the https://with-htpasswd-kuttl-test-sound-panda.apps.rosa.ci-rosa-h-ezi9.l97m.s3.devshift.org/search logger.go:42: 07:27:21 | examples-openshift-with-htpasswd/2-check-unsecured | curl response asserted properly logger.go:42: 07:27:21 | examples-openshift-with-htpasswd/2-check-unsecured | test step completed 2-check-unsecured logger.go:42: 07:27:21 | examples-openshift-with-htpasswd/3-check-unauthorized | starting test step 3-check-unauthorized logger.go:42: 07:27:21 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [./ensure-ingress-host.sh] logger.go:42: 07:27:21 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking the Ingress host value was populated logger.go:42: 07:27:21 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 0 logger.go:42: 07:27:21 | examples-openshift-with-htpasswd/3-check-unauthorized | Hostname is with-htpasswd-kuttl-test-sound-panda.apps.rosa.ci-rosa-h-ezi9.l97m.s3.devshift.org logger.go:42: 07:27:21 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [sh -c JAEGER_USERNAME=wronguser JAEGER_PASSWORD=wrongpassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 07:27:21 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking an expected HTTP response logger.go:42: 07:27:21 | examples-openshift-with-htpasswd/3-check-unauthorized | Running in OpenShift logger.go:42: 07:27:21 | examples-openshift-with-htpasswd/3-check-unauthorized | Using Jaeger basic authentication logger.go:42: 07:27:21 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 1/30 the https://with-htpasswd-kuttl-test-sound-panda.apps.rosa.ci-rosa-h-ezi9.l97m.s3.devshift.org/search logger.go:42: 07:27:22 | examples-openshift-with-htpasswd/3-check-unauthorized | curl response asserted properly logger.go:42: 07:27:22 | examples-openshift-with-htpasswd/3-check-unauthorized | test step completed 3-check-unauthorized logger.go:42: 07:27:22 | examples-openshift-with-htpasswd/4-check-authorized | starting test step 4-check-authorized logger.go:42: 07:27:22 | examples-openshift-with-htpasswd/4-check-authorized | running command: [./ensure-ingress-host.sh] logger.go:42: 07:27:22 | examples-openshift-with-htpasswd/4-check-authorized | Checking the Ingress host value was populated logger.go:42: 07:27:22 | examples-openshift-with-htpasswd/4-check-authorized | Try number 0 logger.go:42: 07:27:22 | examples-openshift-with-htpasswd/4-check-authorized | Hostname is with-htpasswd-kuttl-test-sound-panda.apps.rosa.ci-rosa-h-ezi9.l97m.s3.devshift.org logger.go:42: 07:27:22 | examples-openshift-with-htpasswd/4-check-authorized | running command: [sh -c JAEGER_USERNAME=awesomeuser JAEGER_PASSWORD=awesomepassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE with-htpasswd] logger.go:42: 07:27:22 | examples-openshift-with-htpasswd/4-check-authorized | Checking an expected HTTP response logger.go:42: 07:27:22 | examples-openshift-with-htpasswd/4-check-authorized | Running in OpenShift logger.go:42: 07:27:22 | examples-openshift-with-htpasswd/4-check-authorized | Using Jaeger basic authentication logger.go:42: 07:27:22 | examples-openshift-with-htpasswd/4-check-authorized | Try number 1/30 the https://with-htpasswd-kuttl-test-sound-panda.apps.rosa.ci-rosa-h-ezi9.l97m.s3.devshift.org/search logger.go:42: 07:27:23 | examples-openshift-with-htpasswd/4-check-authorized | curl response asserted properly logger.go:42: 07:27:23 | examples-openshift-with-htpasswd/4-check-authorized | test step completed 4-check-authorized logger.go:42: 07:27:23 | examples-openshift-with-htpasswd | examples-openshift-with-htpasswd events from ns kuttl-test-sound-panda: logger.go:42: 07:27:23 | examples-openshift-with-htpasswd | 2024-09-26 07:27:17 +0000 UTC Normal Pod with-htpasswd-7c58888d86-cw5tf Binding Scheduled Successfully assigned kuttl-test-sound-panda/with-htpasswd-7c58888d86-cw5tf to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:27:23 | examples-openshift-with-htpasswd | 2024-09-26 07:27:17 +0000 UTC Warning Pod with-htpasswd-7c58888d86-cw5tf FailedMount MountVolume.SetUp failed for volume "with-htpasswd-collector-tls-config-volume" : secret "with-htpasswd-collector-headless-tls" not found kubelet logger.go:42: 07:27:23 | examples-openshift-with-htpasswd | 2024-09-26 07:27:17 +0000 UTC Normal ReplicaSet.apps with-htpasswd-7c58888d86 SuccessfulCreate Created pod: with-htpasswd-7c58888d86-cw5tf replicaset-controller logger.go:42: 07:27:23 | examples-openshift-with-htpasswd | 2024-09-26 07:27:17 +0000 UTC Normal Deployment.apps with-htpasswd ScalingReplicaSet Scaled up replica set with-htpasswd-7c58888d86 to 1 deployment-controller logger.go:42: 07:27:23 | examples-openshift-with-htpasswd | 2024-09-26 07:27:18 +0000 UTC Normal Pod with-htpasswd-7c58888d86-cw5tf AddedInterface Add eth0 [10.129.0.84/23] from ovn-kubernetes logger.go:42: 07:27:23 | examples-openshift-with-htpasswd | 2024-09-26 07:27:18 +0000 UTC Normal Pod with-htpasswd-7c58888d86-cw5tf.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:27:23 | examples-openshift-with-htpasswd | 2024-09-26 07:27:18 +0000 UTC Normal Pod with-htpasswd-7c58888d86-cw5tf.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:27:23 | examples-openshift-with-htpasswd | 2024-09-26 07:27:18 +0000 UTC Normal Pod with-htpasswd-7c58888d86-cw5tf.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:27:23 | examples-openshift-with-htpasswd | 2024-09-26 07:27:18 +0000 UTC Normal Pod with-htpasswd-7c58888d86-cw5tf.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:27:23 | examples-openshift-with-htpasswd | 2024-09-26 07:27:18 +0000 UTC Normal Pod with-htpasswd-7c58888d86-cw5tf.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:27:23 | examples-openshift-with-htpasswd | 2024-09-26 07:27:18 +0000 UTC Normal Pod with-htpasswd-7c58888d86-cw5tf.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:27:23 | examples-openshift-with-htpasswd | Deleting namespace: kuttl-test-sound-panda === CONT kuttl/harness/examples-all-in-one-with-options logger.go:42: 07:27:29 | examples-all-in-one-with-options | Creating namespace: kuttl-test-upright-cobra logger.go:42: 07:27:29 | examples-all-in-one-with-options/0-install | starting test step 0-install logger.go:42: 07:27:29 | examples-all-in-one-with-options/0-install | Jaeger:kuttl-test-upright-cobra/my-jaeger created logger.go:42: 07:27:34 | examples-all-in-one-with-options/0-install | test step completed 0-install logger.go:42: 07:27:34 | examples-all-in-one-with-options/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:27:34 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:27:36 | examples-all-in-one-with-options/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:27:43 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443/jaeger MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:27:43 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:27:44 | examples-all-in-one-with-options/1-smoke-test | job.batch/report-span created logger.go:42: 07:27:44 | examples-all-in-one-with-options/1-smoke-test | job.batch/check-span created logger.go:42: 07:27:58 | examples-all-in-one-with-options/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:27:58 | examples-all-in-one-with-options | examples-all-in-one-with-options events from ns kuttl-test-upright-cobra: logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:30 +0000 UTC Normal Pod my-jaeger-5db8478ddc-4p8w9 Binding Scheduled Successfully assigned kuttl-test-upright-cobra/my-jaeger-5db8478ddc-4p8w9 to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:30 +0000 UTC Warning Pod my-jaeger-5db8478ddc-4p8w9 FailedMount MountVolume.SetUp failed for volume "my-jaeger-collector-tls-config-volume" : secret "my-jaeger-collector-headless-tls" not found kubelet logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:30 +0000 UTC Normal ReplicaSet.apps my-jaeger-5db8478ddc SuccessfulCreate Created pod: my-jaeger-5db8478ddc-4p8w9 replicaset-controller logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:30 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-5db8478ddc to 1 deployment-controller logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:31 +0000 UTC Normal Pod my-jaeger-5db8478ddc-4p8w9 AddedInterface Add eth0 [10.128.0.60/23] from ovn-kubernetes logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:31 +0000 UTC Normal Pod my-jaeger-5db8478ddc-4p8w9.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:31 +0000 UTC Normal Pod my-jaeger-5db8478ddc-4p8w9.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:31 +0000 UTC Normal Pod my-jaeger-5db8478ddc-4p8w9.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:31 +0000 UTC Normal Pod my-jaeger-5db8478ddc-4p8w9.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:31 +0000 UTC Normal Pod my-jaeger-5db8478ddc-4p8w9.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:31 +0000 UTC Normal Pod my-jaeger-5db8478ddc-4p8w9.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:36 +0000 UTC Normal Pod my-jaeger-5db8478ddc-4p8w9.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:36 +0000 UTC Normal Pod my-jaeger-5db8478ddc-4p8w9.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:36 +0000 UTC Normal ReplicaSet.apps my-jaeger-5db8478ddc SuccessfulDelete Deleted pod: my-jaeger-5db8478ddc-4p8w9 replicaset-controller logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:36 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-5db8478ddc to 0 from 1 deployment-controller logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:37 +0000 UTC Normal Pod my-jaeger-6b889759-j58hv Binding Scheduled Successfully assigned kuttl-test-upright-cobra/my-jaeger-6b889759-j58hv to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:37 +0000 UTC Normal ReplicaSet.apps my-jaeger-6b889759 SuccessfulCreate Created pod: my-jaeger-6b889759-j58hv replicaset-controller logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:37 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-6b889759 to 1 deployment-controller logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:38 +0000 UTC Normal Pod my-jaeger-6b889759-j58hv AddedInterface Add eth0 [10.128.0.61/23] from ovn-kubernetes logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:38 +0000 UTC Normal Pod my-jaeger-6b889759-j58hv.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:38 +0000 UTC Normal Pod my-jaeger-6b889759-j58hv.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:38 +0000 UTC Normal Pod my-jaeger-6b889759-j58hv.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:38 +0000 UTC Normal Pod my-jaeger-6b889759-j58hv.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:38 +0000 UTC Normal Pod my-jaeger-6b889759-j58hv.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:38 +0000 UTC Normal Pod my-jaeger-6b889759-j58hv.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:44 +0000 UTC Normal Pod check-span-87llg Binding Scheduled Successfully assigned kuttl-test-upright-cobra/check-span-87llg to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:44 +0000 UTC Normal Pod check-span-87llg AddedInterface Add eth0 [10.128.0.62/23] from ovn-kubernetes logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:44 +0000 UTC Normal Pod check-span-87llg.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:44 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-87llg job-controller logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:44 +0000 UTC Normal Pod report-span-j6zhn Binding Scheduled Successfully assigned kuttl-test-upright-cobra/report-span-j6zhn to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:44 +0000 UTC Normal Pod report-span-j6zhn AddedInterface Add eth0 [10.130.0.72/23] from ovn-kubernetes logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:44 +0000 UTC Normal Pod report-span-j6zhn.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:44 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-j6zhn job-controller logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:45 +0000 UTC Normal Pod check-span-87llg.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 829ms (829ms including waiting) kubelet logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:45 +0000 UTC Normal Pod check-span-87llg.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:45 +0000 UTC Normal Pod check-span-87llg.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:45 +0000 UTC Normal Pod report-span-j6zhn.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 891ms (891ms including waiting) kubelet logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:45 +0000 UTC Normal Pod report-span-j6zhn.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:45 +0000 UTC Normal Pod report-span-j6zhn.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:27:58 | examples-all-in-one-with-options | 2024-09-26 07:27:57 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:27:58 | examples-all-in-one-with-options | Deleting namespace: kuttl-test-upright-cobra === CONT kuttl/harness/examples-agent-with-priority-class logger.go:42: 07:28:10 | examples-agent-with-priority-class | Creating namespace: kuttl-test-smart-kit logger.go:42: 07:28:10 | examples-agent-with-priority-class/0-install | starting test step 0-install logger.go:42: 07:28:10 | examples-agent-with-priority-class/0-install | SecurityContextConstraints:/daemonset-with-hostport created logger.go:42: 07:28:10 | examples-agent-with-priority-class/0-install | ServiceAccount:kuttl-test-smart-kit/jaeger-agent-daemonset created logger.go:42: 07:28:10 | examples-agent-with-priority-class/0-install | test step completed 0-install logger.go:42: 07:28:10 | examples-agent-with-priority-class/1-add-policy | starting test step 1-add-policy logger.go:42: 07:28:10 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c oc adm policy --namespace $NAMESPACE add-scc-to-user daemonset-with-hostport -z jaeger-agent-daemonset] logger.go:42: 07:28:11 | examples-agent-with-priority-class/1-add-policy | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:daemonset-with-hostport added: "jaeger-agent-daemonset" logger.go:42: 07:28:11 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c sleep 5] logger.go:42: 07:28:16 | examples-agent-with-priority-class/1-add-policy | test step completed 1-add-policy logger.go:42: 07:28:16 | examples-agent-with-priority-class/2-install | starting test step 2-install logger.go:42: 07:28:16 | examples-agent-with-priority-class/2-install | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE agent-as-daemonset /dev/null] logger.go:42: 07:28:25 | examples-agent-with-priority-class/2-install | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:28:25 | examples-agent-with-priority-class/2-install | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:28:26 | examples-agent-with-priority-class/2-install | job.batch/report-span created logger.go:42: 07:28:26 | examples-agent-with-priority-class/2-install | job.batch/check-span created logger.go:42: 07:28:26 | examples-agent-with-priority-class/2-install | PriorityClass:/high-priority created logger.go:42: 07:28:27 | examples-agent-with-priority-class/2-install | Jaeger:kuttl-test-smart-kit/agent-as-daemonset updated logger.go:42: 07:28:38 | examples-agent-with-priority-class/2-install | test step completed 2-install logger.go:42: 07:28:39 | examples-agent-with-priority-class | examples-agent-with-priority-class events from ns kuttl-test-smart-kit: logger.go:42: 07:28:39 | examples-agent-with-priority-class | 2024-09-26 07:28:19 +0000 UTC Normal Pod agent-as-daemonset-7d48b6c877-84rk7 Binding Scheduled Successfully assigned kuttl-test-smart-kit/agent-as-daemonset-7d48b6c877-84rk7 to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:28:39 | examples-agent-with-priority-class | 2024-09-26 07:28:19 +0000 UTC Normal Pod agent-as-daemonset-7d48b6c877-84rk7 AddedInterface Add eth0 [10.130.0.73/23] from ovn-kubernetes logger.go:42: 07:28:39 | examples-agent-with-priority-class | 2024-09-26 07:28:19 +0000 UTC Normal Pod agent-as-daemonset-7d48b6c877-84rk7.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:28:39 | examples-agent-with-priority-class | 2024-09-26 07:28:19 +0000 UTC Normal Pod agent-as-daemonset-7d48b6c877-84rk7.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:28:39 | examples-agent-with-priority-class | 2024-09-26 07:28:19 +0000 UTC Normal Pod agent-as-daemonset-7d48b6c877-84rk7.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:28:39 | examples-agent-with-priority-class | 2024-09-26 07:28:19 +0000 UTC Normal Pod agent-as-daemonset-7d48b6c877-84rk7.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:28:39 | examples-agent-with-priority-class | 2024-09-26 07:28:19 +0000 UTC Normal Pod agent-as-daemonset-7d48b6c877-84rk7.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:28:39 | examples-agent-with-priority-class | 2024-09-26 07:28:19 +0000 UTC Normal Pod agent-as-daemonset-7d48b6c877-84rk7.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:28:39 | examples-agent-with-priority-class | 2024-09-26 07:28:19 +0000 UTC Normal ReplicaSet.apps agent-as-daemonset-7d48b6c877 SuccessfulCreate Created pod: agent-as-daemonset-7d48b6c877-84rk7 replicaset-controller logger.go:42: 07:28:39 | examples-agent-with-priority-class | 2024-09-26 07:28:19 +0000 UTC Normal Deployment.apps agent-as-daemonset ScalingReplicaSet Scaled up replica set agent-as-daemonset-7d48b6c877 to 1 deployment-controller logger.go:42: 07:28:39 | examples-agent-with-priority-class | 2024-09-26 07:28:26 +0000 UTC Normal Pod check-span-96fhc Binding Scheduled Successfully assigned kuttl-test-smart-kit/check-span-96fhc to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:28:39 | examples-agent-with-priority-class | 2024-09-26 07:28:26 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-96fhc job-controller logger.go:42: 07:28:39 | examples-agent-with-priority-class | 2024-09-26 07:28:26 +0000 UTC Normal Pod report-span-llc5f Binding Scheduled Successfully assigned kuttl-test-smart-kit/report-span-llc5f to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:28:39 | examples-agent-with-priority-class | 2024-09-26 07:28:26 +0000 UTC Normal Pod report-span-llc5f AddedInterface Add eth0 [10.128.0.63/23] from ovn-kubernetes logger.go:42: 07:28:39 | examples-agent-with-priority-class | 2024-09-26 07:28:26 +0000 UTC Normal Pod report-span-llc5f.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:28:39 | examples-agent-with-priority-class | 2024-09-26 07:28:26 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-llc5f job-controller logger.go:42: 07:28:39 | examples-agent-with-priority-class | 2024-09-26 07:28:27 +0000 UTC Warning DaemonSet.apps agent-as-daemonset-agent-daemonset FailedCreate Error creating: pods "agent-as-daemonset-agent-daemonset-" is forbidden: unable to validate against any security context constraint: [provider "anyuid": Forbidden: not usable by user or serviceaccount, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 5775: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 5778: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 6831: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 6832: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 14271: Host ports are not allowed to be used, provider "restricted": Forbidden: not usable by user or serviceaccount, provider "nonroot-v2": Forbidden: not usable by user or serviceaccount, provider "nonroot": Forbidden: not usable by user or serviceaccount, provider "pcap-dedicated-admins": Forbidden: not usable by user or serviceaccount, provider "hostmount-anyuid": Forbidden: not usable by user or serviceaccount, provider "elasticsearch-scc": Forbidden: not usable by user or serviceaccount, provider "daemonset-with-hostport": Forbidden: not usable by user or serviceaccount, provider "hostnetwork-v2": Forbidden: not usable by user or serviceaccount, provider "hostnetwork": Forbidden: not usable by user or serviceaccount, provider "hostaccess": Forbidden: not usable by user or serviceaccount, provider "node-exporter": Forbidden: not usable by user or serviceaccount, provider "privileged": Forbidden: not usable by user or serviceaccount] daemonset-controller logger.go:42: 07:28:39 | examples-agent-with-priority-class | 2024-09-26 07:28:27 +0000 UTC Normal Pod check-span-96fhc AddedInterface Add eth0 [10.129.0.85/23] from ovn-kubernetes logger.go:42: 07:28:39 | examples-agent-with-priority-class | 2024-09-26 07:28:27 +0000 UTC Normal Pod check-span-96fhc.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:28:39 | examples-agent-with-priority-class | 2024-09-26 07:28:27 +0000 UTC Normal Pod check-span-96fhc.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 819ms (819ms including waiting) kubelet logger.go:42: 07:28:39 | examples-agent-with-priority-class | 2024-09-26 07:28:27 +0000 UTC Normal Pod check-span-96fhc.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:28:39 | examples-agent-with-priority-class | 2024-09-26 07:28:27 +0000 UTC Normal Pod check-span-96fhc.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:28:39 | examples-agent-with-priority-class | 2024-09-26 07:28:27 +0000 UTC Normal Pod report-span-llc5f.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 863ms (863ms including waiting) kubelet logger.go:42: 07:28:39 | examples-agent-with-priority-class | 2024-09-26 07:28:27 +0000 UTC Normal Pod report-span-llc5f.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:28:39 | examples-agent-with-priority-class | 2024-09-26 07:28:27 +0000 UTC Normal Pod report-span-llc5f.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:28:39 | examples-agent-with-priority-class | 2024-09-26 07:28:38 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:28:39 | examples-agent-with-priority-class | Deleting namespace: kuttl-test-smart-kit === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (765.06s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.89s) --- PASS: kuttl/harness/examples-simple-prod (69.20s) --- PASS: kuttl/harness/examples-with-sampling (245.71s) --- PASS: kuttl/harness/examples-with-cassandra (105.16s) --- PASS: kuttl/harness/examples-with-badger (37.65s) --- PASS: kuttl/harness/examples-simplest (38.88s) --- PASS: kuttl/harness/examples-simple-prod-with-volumes (66.54s) --- PASS: kuttl/harness/examples-business-application-injected-sidecar (40.40s) --- PASS: kuttl/harness/examples-service-types (54.56s) --- PASS: kuttl/harness/examples-openshift-with-htpasswd (13.20s) --- PASS: kuttl/harness/examples-all-in-one-with-options (41.05s) --- PASS: kuttl/harness/examples-agent-with-priority-class (46.15s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name examples --report --output /logs/artifacts/examples.xml ./artifacts/kuttl-report.xml time="2024-09-26T07:28:56Z" level=debug msg="Setting a new name for the test suites" time="2024-09-26T07:28:56Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-09-26T07:28:56Z" level=debug msg="normalizing test case names" time="2024-09-26T07:28:56Z" level=debug msg="examples/artifacts -> examples_artifacts" time="2024-09-26T07:28:56Z" level=debug msg="examples/examples-simple-prod -> examples_examples_simple_prod" time="2024-09-26T07:28:56Z" level=debug msg="examples/examples-with-sampling -> examples_examples_with_sampling" time="2024-09-26T07:28:56Z" level=debug msg="examples/examples-with-cassandra -> examples_examples_with_cassandra" time="2024-09-26T07:28:56Z" level=debug msg="examples/examples-with-badger -> examples_examples_with_badger" time="2024-09-26T07:28:56Z" level=debug msg="examples/examples-simplest -> examples_examples_simplest" time="2024-09-26T07:28:56Z" level=debug msg="examples/examples-simple-prod-with-volumes -> examples_examples_simple_prod_with_volumes" time="2024-09-26T07:28:56Z" level=debug msg="examples/examples-business-application-injected-sidecar -> examples_examples_business_application_injected_sidecar" time="2024-09-26T07:28:56Z" level=debug msg="examples/examples-service-types -> examples_examples_service_types" time="2024-09-26T07:28:56Z" level=debug msg="examples/examples-openshift-with-htpasswd -> examples_examples_openshift_with_htpasswd" time="2024-09-26T07:28:56Z" level=debug msg="examples/examples-all-in-one-with-options -> examples_examples_all_in_one_with_options" time="2024-09-26T07:28:56Z" level=debug msg="examples/examples-agent-with-priority-class -> examples_examples_agent_with_priority_class" +---------------------------------------------------------+--------+ | NAME | RESULT | +---------------------------------------------------------+--------+ | examples_artifacts | passed | | examples_examples_simple_prod | passed | | examples_examples_with_sampling | passed | | examples_examples_with_cassandra | passed | | examples_examples_with_badger | passed | | examples_examples_simplest | passed | | examples_examples_simple_prod_with_volumes | passed | | examples_examples_business_application_injected_sidecar | passed | | examples_examples_service_types | passed | | examples_examples_openshift_with_htpasswd | passed | | examples_examples_all_in_one_with_options | passed | | examples_examples_agent_with_priority_class | passed | +---------------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh generate false true + '[' 3 -ne 3 ']' + test_suite_name=generate + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/generate.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-generate make[2]: Entering directory '/tmp/jaeger-tests' test -s /tmp/jaeger-tests/bin/operator-sdk || curl -sLo /tmp/jaeger-tests/bin/operator-sdk https://github.com/operator-framework/operator-sdk/releases/download/v1.32.0/operator-sdk_`go env GOOS`_`go env GOARCH` ./hack/install/install-golangci-lint.sh Installing golangci-lint golangci-lint 1.55.2 is installed already ./hack/install/install-goimports.sh Installing goimports Try 0... go install golang.org/x/tools/cmd/goimports@v0.1.12 >>>> Formatting code... ./.ci/format.sh >>>> Building... ./hack/install/install-dependencies.sh Installing go dependencies Try 0... go mod download GOOS= GOARCH= CGO_ENABLED=0 GO111MODULE=on go build -ldflags "-X "github.com/jaegertracing/jaeger-operator/pkg/version".version="1.61.0" -X "github.com/jaegertracing/jaeger-operator/pkg/version".buildDate=2024-09-26T07:28:58Z -X "github.com/jaegertracing/jaeger-operator/pkg/version".defaultJaeger="1.61.0"" -o "bin/jaeger-operator" main.go JAEGER_VERSION="1.61.0" ./tests/e2e/generate/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.32 True False 42m Cluster version is 4.15.32' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.32 True False 42m Cluster version is 4.15.32' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/generate/render.sh ++ export SUITE_DIR=./tests/e2e/generate ++ SUITE_DIR=./tests/e2e/generate ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/generate ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + '[' true = true ']' + skip_test generate 'This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 2 -ne 2 ']' + test_name=generate + message='This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/generate/_build + '[' _build '!=' _build ']' + rm -rf generate + warning 'generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed\e[0m' WAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running generate E2E tests' Running generate E2E tests + cd tests/e2e/generate/_build + set +e + KUBECONFIG=/tmp/kubeconfig-488946273 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-rosa-h-ezi9.l97m.s3.devshift.org:443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 07:29:15 | artifacts | Creating namespace: kuttl-test-resolved-albacore logger.go:42: 07:29:15 | artifacts | artifacts events from ns kuttl-test-resolved-albacore: logger.go:42: 07:29:15 | artifacts | Deleting namespace: kuttl-test-resolved-albacore === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (6.51s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.90s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name generate --report --output /logs/artifacts/generate.xml ./artifacts/kuttl-report.xml time="2024-09-26T07:29:21Z" level=debug msg="Setting a new name for the test suites" time="2024-09-26T07:29:21Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-09-26T07:29:21Z" level=debug msg="normalizing test case names" time="2024-09-26T07:29:21Z" level=debug msg="generate/artifacts -> generate_artifacts" +--------------------+--------+ | NAME | RESULT | +--------------------+--------+ | generate_artifacts | passed | +--------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh upgrade false true + '[' 3 -ne 3 ']' + test_suite_name=upgrade + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/upgrade.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-upgrade make[2]: Entering directory '/tmp/jaeger-tests' make docker JAEGER_VERSION=1.61.1 IMG="quay.io//jaeger-operator:next" make[3]: Entering directory '/tmp/jaeger-tests' [ ! -z "true" ] || docker build --build-arg=GOPROXY= --build-arg=VERSION="1.61.0" --build-arg=JAEGER_VERSION=1.61.1 --build-arg=TARGETARCH= --build-arg VERSION_DATE=2024-09-26T07:29:21Z --build-arg VERSION_PKG="github.com/jaegertracing/jaeger-operator/pkg/version" -t "quay.io//jaeger-operator:next" . make[3]: Leaving directory '/tmp/jaeger-tests' touch build-e2e-upgrade-image SKIP_ES_EXTERNAL=true IMG=quay.io//jaeger-operator:"1.61.0" JAEGER_OPERATOR_VERSION="1.61.0" JAEGER_VERSION="1.61.0" ./tests/e2e/upgrade/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.32 True False 42m Cluster version is 4.15.32' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.32 True False 42m Cluster version is 4.15.32' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/upgrade/render.sh ++ export SUITE_DIR=./tests/e2e/upgrade ++ SUITE_DIR=./tests/e2e/upgrade ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/upgrade ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + export JAEGER_NAME + '[' true = true ']' + skip_test upgrade 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade + warning 'upgrade: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade: Test not supported in OpenShift\e[0m' WAR: upgrade: Test not supported in OpenShift + '[' true = true ']' + skip_test upgrade-from-latest-release 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade-from-latest-release + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade-from-latest-release + warning 'upgrade-from-latest-release: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade-from-latest-release: Test not supported in OpenShift\e[0m' WAR: upgrade-from-latest-release: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running upgrade E2E tests' Running upgrade E2E tests + cd tests/e2e/upgrade/_build + set +e + KUBECONFIG=/tmp/kubeconfig-488946273 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-rosa-h-ezi9.l97m.s3.devshift.org:443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 07:29:23 | artifacts | Creating namespace: kuttl-test-measured-chimp logger.go:42: 07:29:23 | artifacts | artifacts events from ns kuttl-test-measured-chimp: logger.go:42: 07:29:23 | artifacts | Deleting namespace: kuttl-test-measured-chimp === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (6.48s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.89s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name upgrade --report --output /logs/artifacts/upgrade.xml ./artifacts/kuttl-report.xml time="2024-09-26T07:29:29Z" level=debug msg="Setting a new name for the test suites" time="2024-09-26T07:29:29Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-09-26T07:29:29Z" level=debug msg="normalizing test case names" time="2024-09-26T07:29:29Z" level=debug msg="upgrade/artifacts -> upgrade_artifacts" +-------------------+--------+ | NAME | RESULT | +-------------------+--------+ | upgrade_artifacts | passed | +-------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh sidecar false true + '[' 3 -ne 3 ']' + test_suite_name=sidecar + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/sidecar.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-sidecar make[2]: Entering directory '/tmp/jaeger-tests' ./tests/e2e/sidecar/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.32 True False 42m Cluster version is 4.15.32' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.32 True False 42m Cluster version is 4.15.32' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/sidecar/render.sh ++ export SUITE_DIR=./tests/e2e/sidecar ++ SUITE_DIR=./tests/e2e/sidecar ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/sidecar ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + jaeger_service_name=order + start_test sidecar-deployment + '[' 1 -ne 1 ']' + test_name=sidecar-deployment + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-deployment' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-deployment\e[0m' Rendering files for test sidecar-deployment + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build + '[' _build '!=' _build ']' + mkdir -p sidecar-deployment + cd sidecar-deployment + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-namespace + '[' 1 -ne 1 ']' + test_name=sidecar-namespace + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-namespace' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-namespace\e[0m' Rendering files for test sidecar-namespace + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-deployment + '[' sidecar-deployment '!=' _build ']' + cd .. + mkdir -p sidecar-namespace + cd sidecar-namespace + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-skip-webhook + '[' 1 -ne 1 ']' + test_name=sidecar-skip-webhook + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-skip-webhook' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-skip-webhook\e[0m' Rendering files for test sidecar-skip-webhook + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-namespace + '[' sidecar-namespace '!=' _build ']' + cd .. + mkdir -p sidecar-skip-webhook + cd sidecar-skip-webhook + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running sidecar E2E tests' Running sidecar E2E tests + cd tests/e2e/sidecar/_build + set +e + KUBECONFIG=/tmp/kubeconfig-488946273 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-rosa-h-ezi9.l97m.s3.devshift.org:443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 4 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/sidecar-deployment === PAUSE kuttl/harness/sidecar-deployment === RUN kuttl/harness/sidecar-namespace === PAUSE kuttl/harness/sidecar-namespace === RUN kuttl/harness/sidecar-skip-webhook === PAUSE kuttl/harness/sidecar-skip-webhook === CONT kuttl/harness/artifacts logger.go:42: 07:29:38 | artifacts | Creating namespace: kuttl-test-allowed-oyster logger.go:42: 07:29:38 | artifacts | artifacts events from ns kuttl-test-allowed-oyster: logger.go:42: 07:29:38 | artifacts | Deleting namespace: kuttl-test-allowed-oyster === CONT kuttl/harness/sidecar-namespace logger.go:42: 07:29:44 | sidecar-namespace | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:29:44 | sidecar-namespace | Creating namespace: kuttl-test-apparent-wombat logger.go:42: 07:29:44 | sidecar-namespace/0-install | starting test step 0-install case.go:364: failed in step 0-install case.go:366: Internal error occurred: failed calling webhook "mjaeger.kb.io": failed to call webhook: Post "https://jaeger-operator-service.openshift-distributed-tracing.svc:443/mutate-jaegertracing-io-v1-jaeger?timeout=10s": context deadline exceeded logger.go:42: 07:29:54 | sidecar-namespace | sidecar-namespace events from ns kuttl-test-apparent-wombat: logger.go:42: 07:29:54 | sidecar-namespace | Deleting namespace: kuttl-test-apparent-wombat === CONT kuttl/harness/sidecar-skip-webhook logger.go:42: 07:30:00 | sidecar-skip-webhook | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:30:00 | sidecar-skip-webhook | Creating namespace: kuttl-test-model-baboon logger.go:42: 07:30:00 | sidecar-skip-webhook/0-install | starting test step 0-install logger.go:42: 07:30:01 | sidecar-skip-webhook/0-install | Jaeger:kuttl-test-model-baboon/agent-as-sidecar created logger.go:42: 07:30:04 | sidecar-skip-webhook/0-install | test step completed 0-install logger.go:42: 07:30:04 | sidecar-skip-webhook/1-install | starting test step 1-install logger.go:42: 07:30:04 | sidecar-skip-webhook/1-install | Deployment:kuttl-test-model-baboon/vertx-create-span-sidecar created logger.go:42: 07:30:12 | sidecar-skip-webhook/1-install | test step completed 1-install logger.go:42: 07:30:12 | sidecar-skip-webhook/2-add-anotation-and-label | starting test step 2-add-anotation-and-label logger.go:42: 07:30:12 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name=jaeger-operator --namespace kuttl-test-model-baboon] logger.go:42: 07:30:12 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar labeled logger.go:42: 07:30:12 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-model-baboon] logger.go:42: 07:30:13 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 07:30:13 | sidecar-skip-webhook/2-add-anotation-and-label | test step completed 2-add-anotation-and-label logger.go:42: 07:30:13 | sidecar-skip-webhook/3-remove-label | starting test step 3-remove-label logger.go:42: 07:30:13 | sidecar-skip-webhook/3-remove-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name- --namespace kuttl-test-model-baboon] logger.go:42: 07:30:13 | sidecar-skip-webhook/3-remove-label | deployment.apps/vertx-create-span-sidecar unlabeled logger.go:42: 07:30:15 | sidecar-skip-webhook/3-remove-label | test step completed 3-remove-label logger.go:42: 07:30:15 | sidecar-skip-webhook | sidecar-skip-webhook events from ns kuttl-test-model-baboon: logger.go:42: 07:30:15 | sidecar-skip-webhook | 2024-09-26 07:30:01 +0000 UTC Normal Pod agent-as-sidecar-84bbfb8bd9-l72qj Binding Scheduled Successfully assigned kuttl-test-model-baboon/agent-as-sidecar-84bbfb8bd9-l72qj to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:30:15 | sidecar-skip-webhook | 2024-09-26 07:30:01 +0000 UTC Normal Pod agent-as-sidecar-84bbfb8bd9-l72qj AddedInterface Add eth0 [10.128.0.64/23] from ovn-kubernetes logger.go:42: 07:30:15 | sidecar-skip-webhook | 2024-09-26 07:30:01 +0000 UTC Normal Pod agent-as-sidecar-84bbfb8bd9-l72qj.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:30:15 | sidecar-skip-webhook | 2024-09-26 07:30:01 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-84bbfb8bd9 SuccessfulCreate Created pod: agent-as-sidecar-84bbfb8bd9-l72qj replicaset-controller logger.go:42: 07:30:15 | sidecar-skip-webhook | 2024-09-26 07:30:01 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-84bbfb8bd9 to 1 deployment-controller logger.go:42: 07:30:15 | sidecar-skip-webhook | 2024-09-26 07:30:02 +0000 UTC Normal Pod agent-as-sidecar-84bbfb8bd9-l72qj.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:30:15 | sidecar-skip-webhook | 2024-09-26 07:30:02 +0000 UTC Normal Pod agent-as-sidecar-84bbfb8bd9-l72qj.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:30:15 | sidecar-skip-webhook | 2024-09-26 07:30:04 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-29m4p Binding Scheduled Successfully assigned kuttl-test-model-baboon/vertx-create-span-sidecar-84d458b68c-29m4p to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:30:15 | sidecar-skip-webhook | 2024-09-26 07:30:04 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-29m4p replicaset-controller logger.go:42: 07:30:15 | sidecar-skip-webhook | 2024-09-26 07:30:04 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 07:30:15 | sidecar-skip-webhook | 2024-09-26 07:30:05 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-29m4p AddedInterface Add eth0 [10.129.0.86/23] from ovn-kubernetes logger.go:42: 07:30:15 | sidecar-skip-webhook | 2024-09-26 07:30:05 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-29m4p.spec.containers{vertx-create-span-sidecar} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 07:30:15 | sidecar-skip-webhook | 2024-09-26 07:30:11 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-29m4p.spec.containers{vertx-create-span-sidecar} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 6.459s (6.459s including waiting) kubelet logger.go:42: 07:30:15 | sidecar-skip-webhook | 2024-09-26 07:30:11 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-29m4p.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:30:15 | sidecar-skip-webhook | 2024-09-26 07:30:11 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-29m4p.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:30:15 | sidecar-skip-webhook | 2024-09-26 07:30:13 +0000 UTC Normal Pod vertx-create-span-sidecar-9cc7fdc6c-lw2nr Binding Scheduled Successfully assigned kuttl-test-model-baboon/vertx-create-span-sidecar-9cc7fdc6c-lw2nr to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:30:15 | sidecar-skip-webhook | 2024-09-26 07:30:13 +0000 UTC Normal Pod vertx-create-span-sidecar-9cc7fdc6c-lw2nr AddedInterface Add eth0 [10.128.0.65/23] from ovn-kubernetes logger.go:42: 07:30:15 | sidecar-skip-webhook | 2024-09-26 07:30:13 +0000 UTC Normal Pod vertx-create-span-sidecar-9cc7fdc6c-lw2nr.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:30:15 | sidecar-skip-webhook | 2024-09-26 07:30:13 +0000 UTC Normal Pod vertx-create-span-sidecar-9cc7fdc6c-lw2nr.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:30:15 | sidecar-skip-webhook | 2024-09-26 07:30:13 +0000 UTC Normal Pod vertx-create-span-sidecar-9cc7fdc6c-lw2nr.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:30:15 | sidecar-skip-webhook | 2024-09-26 07:30:13 +0000 UTC Normal Pod vertx-create-span-sidecar-9cc7fdc6c-lw2nr.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:30:15 | sidecar-skip-webhook | 2024-09-26 07:30:13 +0000 UTC Normal Pod vertx-create-span-sidecar-9cc7fdc6c-lw2nr.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:30:15 | sidecar-skip-webhook | 2024-09-26 07:30:13 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-9cc7fdc6c SuccessfulCreate Created pod: vertx-create-span-sidecar-9cc7fdc6c-lw2nr replicaset-controller logger.go:42: 07:30:15 | sidecar-skip-webhook | 2024-09-26 07:30:13 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-9cc7fdc6c to 1 deployment-controller logger.go:42: 07:30:15 | sidecar-skip-webhook | 2024-09-26 07:30:14 +0000 UTC Normal Pod vertx-create-span-sidecar-9cc7fdc6c-lw2nr.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:30:15 | sidecar-skip-webhook | Deleting namespace: kuttl-test-model-baboon === CONT kuttl/harness/sidecar-deployment logger.go:42: 07:30:21 | sidecar-deployment | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:30:21 | sidecar-deployment | Creating namespace: kuttl-test-honest-gopher logger.go:42: 07:30:21 | sidecar-deployment/0-install | starting test step 0-install logger.go:42: 07:30:21 | sidecar-deployment/0-install | Jaeger:kuttl-test-honest-gopher/agent-as-sidecar created logger.go:42: 07:30:24 | sidecar-deployment/0-install | test step completed 0-install logger.go:42: 07:30:24 | sidecar-deployment/1-install | starting test step 1-install logger.go:42: 07:30:24 | sidecar-deployment/1-install | Deployment:kuttl-test-honest-gopher/vertx-create-span-sidecar created logger.go:42: 07:30:26 | sidecar-deployment/1-install | test step completed 1-install logger.go:42: 07:30:26 | sidecar-deployment/2-enable-injection | starting test step 2-enable-injection logger.go:42: 07:30:26 | sidecar-deployment/2-enable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-honest-gopher] logger.go:42: 07:30:27 | sidecar-deployment/2-enable-injection | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 07:30:34 | sidecar-deployment/2-enable-injection | test step completed 2-enable-injection logger.go:42: 07:30:34 | sidecar-deployment/3-find-service | starting test step 3-find-service logger.go:42: 07:30:34 | sidecar-deployment/3-find-service | Job:kuttl-test-honest-gopher/00-find-service created logger.go:42: 07:30:47 | sidecar-deployment/3-find-service | test step completed 3-find-service logger.go:42: 07:30:47 | sidecar-deployment/4-other-instance | starting test step 4-other-instance logger.go:42: 07:30:47 | sidecar-deployment/4-other-instance | Jaeger:kuttl-test-honest-gopher/agent-as-sidecar2 created logger.go:42: 07:30:52 | sidecar-deployment/4-other-instance | test step completed 4-other-instance logger.go:42: 07:30:52 | sidecar-deployment/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 07:30:52 | sidecar-deployment/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 07:30:52 | sidecar-deployment/6-find-service | starting test step 6-find-service logger.go:42: 07:30:53 | sidecar-deployment/6-find-service | Job:kuttl-test-honest-gopher/01-find-service created logger.go:42: 07:31:05 | sidecar-deployment/6-find-service | test step completed 6-find-service logger.go:42: 07:31:05 | sidecar-deployment/7-disable-injection | starting test step 7-disable-injection logger.go:42: 07:31:05 | sidecar-deployment/7-disable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=false --namespace kuttl-test-honest-gopher] logger.go:42: 07:31:13 | sidecar-deployment/7-disable-injection | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 07:31:16 | sidecar-deployment/7-disable-injection | test step completed 7-disable-injection logger.go:42: 07:31:16 | sidecar-deployment | sidecar-deployment events from ns kuttl-test-honest-gopher: logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:22 +0000 UTC Normal Pod agent-as-sidecar-59fb67f4d7-2tmkr Binding Scheduled Successfully assigned kuttl-test-honest-gopher/agent-as-sidecar-59fb67f4d7-2tmkr to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:22 +0000 UTC Normal Pod agent-as-sidecar-59fb67f4d7-2tmkr AddedInterface Add eth0 [10.128.0.66/23] from ovn-kubernetes logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:22 +0000 UTC Normal Pod agent-as-sidecar-59fb67f4d7-2tmkr.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:22 +0000 UTC Normal Pod agent-as-sidecar-59fb67f4d7-2tmkr.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:22 +0000 UTC Normal Pod agent-as-sidecar-59fb67f4d7-2tmkr.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:22 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-59fb67f4d7 SuccessfulCreate Created pod: agent-as-sidecar-59fb67f4d7-2tmkr replicaset-controller logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:22 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-59fb67f4d7 to 1 deployment-controller logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:24 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-d9mlt Binding Scheduled Successfully assigned kuttl-test-honest-gopher/vertx-create-span-sidecar-84d458b68c-d9mlt to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:24 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-d9mlt AddedInterface Add eth0 [10.129.0.87/23] from ovn-kubernetes logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:24 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-d9mlt.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:24 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-d9mlt replicaset-controller logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:24 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:25 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-d9mlt.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:25 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-d9mlt.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:27 +0000 UTC Normal Pod vertx-create-span-sidecar-69c94f45d7-qgff8 Binding Scheduled Successfully assigned kuttl-test-honest-gopher/vertx-create-span-sidecar-69c94f45d7-qgff8 to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:27 +0000 UTC Normal Pod vertx-create-span-sidecar-69c94f45d7-qgff8 AddedInterface Add eth0 [10.130.0.75/23] from ovn-kubernetes logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:27 +0000 UTC Normal Pod vertx-create-span-sidecar-69c94f45d7-qgff8.spec.containers{vertx-create-span-sidecar} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:27 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-69c94f45d7 SuccessfulCreate Created pod: vertx-create-span-sidecar-69c94f45d7-qgff8 replicaset-controller logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:27 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-69c94f45d7 to 1 deployment-controller logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:32 +0000 UTC Normal Pod vertx-create-span-sidecar-69c94f45d7-qgff8.spec.containers{vertx-create-span-sidecar} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 5.147s (5.147s including waiting) kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:32 +0000 UTC Normal Pod vertx-create-span-sidecar-69c94f45d7-qgff8.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:32 +0000 UTC Normal Pod vertx-create-span-sidecar-69c94f45d7-qgff8.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:32 +0000 UTC Normal Pod vertx-create-span-sidecar-69c94f45d7-qgff8.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:33 +0000 UTC Normal Pod vertx-create-span-sidecar-69c94f45d7-qgff8.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:33 +0000 UTC Normal Pod vertx-create-span-sidecar-69c94f45d7-qgff8.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:33 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-d9mlt.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.0.87:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:33 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-d9mlt.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.0.87:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:34 +0000 UTC Normal Pod 00-find-service-n9sv2 Binding Scheduled Successfully assigned kuttl-test-honest-gopher/00-find-service-n9sv2 to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:34 +0000 UTC Normal Pod 00-find-service-n9sv2 AddedInterface Add eth0 [10.128.0.67/23] from ovn-kubernetes logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:34 +0000 UTC Normal Pod 00-find-service-n9sv2.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:34 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-n9sv2 job-controller logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:35 +0000 UTC Normal Pod 00-find-service-n9sv2.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 861ms (861ms including waiting) kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:35 +0000 UTC Normal Pod 00-find-service-n9sv2.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:35 +0000 UTC Normal Pod 00-find-service-n9sv2.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:36 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-d9mlt.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:36 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-d9mlt.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.0.87:8080/": read tcp 10.129.0.2:47318->10.129.0.87:8080: read: connection reset by peer kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:36 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-d9mlt.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.0.87:8080/": dial tcp 10.129.0.87:8080: connect: connection refused kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:40 +0000 UTC Warning Pod vertx-create-span-sidecar-69c94f45d7-qgff8.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.75:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:40 +0000 UTC Warning Pod vertx-create-span-sidecar-69c94f45d7-qgff8.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.130.0.75:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:42 +0000 UTC Normal Pod vertx-create-span-sidecar-69c94f45d7-qgff8.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:43 +0000 UTC Warning Pod vertx-create-span-sidecar-69c94f45d7-qgff8.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.75:8080/": read tcp 10.130.0.2:55502->10.130.0.75:8080: read: connection reset by peer kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:43 +0000 UTC Warning Pod vertx-create-span-sidecar-69c94f45d7-qgff8.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.75:8080/": dial tcp 10.130.0.75:8080: connect: connection refused kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:43 +0000 UTC Normal Pod vertx-create-span-sidecar-69c94f45d7-qgff8.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:47 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:47 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-d9mlt.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.0.87:8080/": read tcp 10.129.0.2:43930->10.129.0.87:8080: read: connection reset by peer kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:48 +0000 UTC Normal Pod agent-as-sidecar2-79bddb6cbc-n695g Binding Scheduled Successfully assigned kuttl-test-honest-gopher/agent-as-sidecar2-79bddb6cbc-n695g to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:48 +0000 UTC Warning Pod agent-as-sidecar2-79bddb6cbc-n695g FailedMount MountVolume.SetUp failed for volume "agent-as-sidecar2-collector-tls-config-volume" : secret "agent-as-sidecar2-collector-headless-tls" not found kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:48 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-79bddb6cbc SuccessfulCreate Created pod: agent-as-sidecar2-79bddb6cbc-n695g replicaset-controller logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:48 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-79bddb6cbc to 1 deployment-controller logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:49 +0000 UTC Normal Pod agent-as-sidecar2-79bddb6cbc-n695g AddedInterface Add eth0 [10.129.0.88/23] from ovn-kubernetes logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:49 +0000 UTC Normal Pod agent-as-sidecar2-79bddb6cbc-n695g.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:49 +0000 UTC Normal Pod agent-as-sidecar2-79bddb6cbc-n695g.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:49 +0000 UTC Normal Pod agent-as-sidecar2-79bddb6cbc-n695g.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:52 +0000 UTC Normal Pod agent-as-sidecar-59fb67f4d7-2tmkr.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:53 +0000 UTC Normal Pod 01-find-service-b72c4 Binding Scheduled Successfully assigned kuttl-test-honest-gopher/01-find-service-b72c4 to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:53 +0000 UTC Normal Pod 01-find-service-b72c4 AddedInterface Add eth0 [10.128.0.68/23] from ovn-kubernetes logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:53 +0000 UTC Normal Pod 01-find-service-b72c4.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:53 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-b72c4 job-controller logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:54 +0000 UTC Normal Pod 01-find-service-b72c4.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 854ms (854ms including waiting) kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:54 +0000 UTC Normal Pod 01-find-service-b72c4.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:54 +0000 UTC Normal Pod 01-find-service-b72c4.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:54 +0000 UTC Normal Pod vertx-create-span-sidecar-678bcfdb75-hxrqj Binding Scheduled Successfully assigned kuttl-test-honest-gopher/vertx-create-span-sidecar-678bcfdb75-hxrqj to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:54 +0000 UTC Normal Pod vertx-create-span-sidecar-678bcfdb75-hxrqj AddedInterface Add eth0 [10.128.0.69/23] from ovn-kubernetes logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:54 +0000 UTC Normal Pod vertx-create-span-sidecar-678bcfdb75-hxrqj.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:54 +0000 UTC Normal Pod vertx-create-span-sidecar-678bcfdb75-hxrqj.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:54 +0000 UTC Normal Pod vertx-create-span-sidecar-678bcfdb75-hxrqj.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:54 +0000 UTC Normal Pod vertx-create-span-sidecar-678bcfdb75-hxrqj.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:54 +0000 UTC Normal Pod vertx-create-span-sidecar-678bcfdb75-hxrqj.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:54 +0000 UTC Normal Pod vertx-create-span-sidecar-678bcfdb75-hxrqj.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:54 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-678bcfdb75 SuccessfulCreate Created pod: vertx-create-span-sidecar-678bcfdb75-hxrqj replicaset-controller logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:54 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulDelete Deleted pod: vertx-create-span-sidecar-84d458b68c-d9mlt replicaset-controller logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:54 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-84d458b68c to 0 from 1 deployment-controller logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:30:54 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-678bcfdb75 to 1 from 0 deployment-controller logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:31:02 +0000 UTC Warning Pod vertx-create-span-sidecar-678bcfdb75-hxrqj.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.128.0.69:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:31:02 +0000 UTC Warning Pod vertx-create-span-sidecar-678bcfdb75-hxrqj.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.0.69:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:31:04 +0000 UTC Normal Pod vertx-create-span-sidecar-678bcfdb75-hxrqj.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:31:05 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:31:05 +0000 UTC Warning Pod vertx-create-span-sidecar-678bcfdb75-hxrqj.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.0.69:8080/": read tcp 10.128.0.2:38980->10.128.0.69:8080: read: connection reset by peer kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:31:05 +0000 UTC Warning Pod vertx-create-span-sidecar-678bcfdb75-hxrqj.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.0.69:8080/": dial tcp 10.128.0.69:8080: connect: connection refused kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:31:13 +0000 UTC Normal Pod vertx-create-span-sidecar-658985b9b5-mblts Binding Scheduled Successfully assigned kuttl-test-honest-gopher/vertx-create-span-sidecar-658985b9b5-mblts to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:31:13 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-658985b9b5 SuccessfulCreate Created pod: vertx-create-span-sidecar-658985b9b5-mblts replicaset-controller logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:31:13 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-69c94f45d7 SuccessfulDelete Deleted pod: vertx-create-span-sidecar-69c94f45d7-qgff8 replicaset-controller logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:31:13 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-69c94f45d7 to 0 from 1 deployment-controller logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:31:13 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-658985b9b5 to 1 from 0 deployment-controller logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:31:14 +0000 UTC Normal Pod vertx-create-span-sidecar-658985b9b5-mblts AddedInterface Add eth0 [10.129.0.89/23] from ovn-kubernetes logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:31:14 +0000 UTC Normal Pod vertx-create-span-sidecar-658985b9b5-mblts.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:31:14 +0000 UTC Normal Pod vertx-create-span-sidecar-658985b9b5-mblts.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:31:16 | sidecar-deployment | 2024-09-26 07:31:14 +0000 UTC Normal Pod vertx-create-span-sidecar-658985b9b5-mblts.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:31:16 | sidecar-deployment | Deleting namespace: kuttl-test-honest-gopher === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- FAIL: kuttl (105.50s) --- FAIL: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.91s) --- FAIL: kuttl/harness/sidecar-namespace (16.29s) --- PASS: kuttl/harness/sidecar-skip-webhook (20.97s) --- PASS: kuttl/harness/sidecar-deployment (61.68s) FAIL + exit_code=1 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name sidecar --report --output /logs/artifacts/sidecar.xml ./artifacts/kuttl-report.xml time="2024-09-26T07:31:23Z" level=debug msg="Setting a new name for the test suites" time="2024-09-26T07:31:23Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-09-26T07:31:23Z" level=debug msg="normalizing test case names" time="2024-09-26T07:31:23Z" level=debug msg="sidecar/artifacts -> sidecar_artifacts" time="2024-09-26T07:31:23Z" level=debug msg="sidecar/sidecar-namespace -> sidecar_sidecar_namespace" time="2024-09-26T07:31:23Z" level=debug msg="sidecar/sidecar-skip-webhook -> sidecar_sidecar_skip_webhook" time="2024-09-26T07:31:23Z" level=debug msg="sidecar/sidecar-deployment -> sidecar_sidecar_deployment" +------------------------------+--------+ | NAME | RESULT | +------------------------------+--------+ | sidecar_artifacts | passed | | sidecar_sidecar_namespace | failed | | sidecar_sidecar_skip_webhook | passed | | sidecar_sidecar_deployment | passed | +------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh streaming false true + '[' 3 -ne 3 ']' + test_suite_name=streaming + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/streaming.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-streaming make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ SKIP_ES_EXTERNAL=true \ ./tests/e2e/streaming/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.32 True False 44m Cluster version is 4.15.32' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.32 True False 44m Cluster version is 4.15.32' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/streaming/render.sh ++ export SUITE_DIR=./tests/e2e/streaming ++ SUITE_DIR=./tests/e2e/streaming ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/streaming ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + '[' false = true ']' + start_test streaming-simple + '[' 1 -ne 1 ']' + test_name=streaming-simple + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-simple' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-simple\e[0m' Rendering files for test streaming-simple + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build + '[' _build '!=' _build ']' + mkdir -p streaming-simple + cd streaming-simple + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/streaming-jaeger-assert.yaml.template -o ./04-assert.yaml + render_smoke_test simple-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=simple-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + export JAEGER_NAME=simple-streaming + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-tls + '[' 1 -ne 1 ']' + test_name=streaming-with-tls + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-tls' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-tls\e[0m' Rendering files for test streaming-with-tls + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-simple + '[' streaming-simple '!=' _build ']' + cd .. + mkdir -p streaming-with-tls + cd streaming-with-tls + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + render_smoke_test tls-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=tls-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + export JAEGER_NAME=tls-streaming + JAEGER_NAME=tls-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-autoprovisioning-autoscale + '[' 1 -ne 1 ']' + test_name=streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-autoprovisioning-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-autoprovisioning-autoscale\e[0m' Rendering files for test streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-with-tls + '[' streaming-with-tls '!=' _build ']' + cd .. + mkdir -p streaming-with-autoprovisioning-autoscale + cd streaming-with-autoprovisioning-autoscale + '[' true = true ']' + rm ./00-install.yaml ./00-assert.yaml + render_install_elasticsearch upstream 01 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=01 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./01-assert.yaml + jaeger_name=auto-provisioned + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="20Mi"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="500m"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.autoscale=true ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.minReplicas=1 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.maxReplicas=2 ./02-install.yaml + render_assert_kafka true auto-provisioned 03 + '[' 3 -ne 3 ']' + autoprovisioned=true + cluster_name=auto-provisioned + test_step=03 + '[' true = true ']' + is_kafka_minimal_enabled + namespaces=(observability openshift-operators openshift-distributed-tracing) + for i in "${namespaces[@]}" ++ kubectl get pods -n observability -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-operators -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-distributed-tracing -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled=true + '[' true == true ']' + return 0 + replicas=1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./03-assert.yaml ++ expr 03 + 1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./04-assert.yaml ++ expr 03 + 2 + CLUSTER_NAME=auto-provisioned + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./05-assert.yaml + version_lt 1.30 1.23 ++ echo 1.30 1.23 ++ tr ' ' '\n' ++ sort -rV ++ head -n 1 + test 1.30 '!=' 1.30 + rm ./08-assert.yaml + skip_test streaming-with-tls 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=streaming-with-tls + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-with-autoprovisioning-autoscale + '[' streaming-with-autoprovisioning-autoscale '!=' _build ']' + cd .. + rm -rf streaming-with-tls + warning 'streaming-with-tls: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: streaming-with-tls: This test is flaky in Prow CI\e[0m' WAR: streaming-with-tls: This test is flaky in Prow CI + skip_test streaming-simple 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=streaming-simple + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build + '[' _build '!=' _build ']' + rm -rf streaming-simple + warning 'streaming-simple: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: streaming-simple: This test is flaky in Prow CI\e[0m' WAR: streaming-simple: This test is flaky in Prow CI make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running streaming E2E tests' Running streaming E2E tests + cd tests/e2e/streaming/_build + set +e + KUBECONFIG=/tmp/kubeconfig-488946273 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-rosa-h-ezi9.l97m.s3.devshift.org:443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 2 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/streaming-with-autoprovisioning-autoscale === PAUSE kuttl/harness/streaming-with-autoprovisioning-autoscale === CONT kuttl/harness/artifacts logger.go:42: 07:31:37 | artifacts | Creating namespace: kuttl-test-prepared-kingfish logger.go:42: 07:31:37 | artifacts | artifacts events from ns kuttl-test-prepared-kingfish: logger.go:42: 07:31:37 | artifacts | Deleting namespace: kuttl-test-prepared-kingfish === CONT kuttl/harness/streaming-with-autoprovisioning-autoscale logger.go:42: 07:31:43 | streaming-with-autoprovisioning-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:31:43 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_0.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:31:43 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_1.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:31:43 | streaming-with-autoprovisioning-autoscale | Creating namespace: kuttl-test-grateful-goldfish logger.go:42: 07:31:43 | streaming-with-autoprovisioning-autoscale/1-install | starting test step 1-install logger.go:42: 07:31:43 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc create sa deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 07:31:43 | streaming-with-autoprovisioning-autoscale/1-install | serviceaccount/deploy-elasticsearch created logger.go:42: 07:31:43 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc adm policy add-scc-to-user privileged -z deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 07:31:43 | streaming-with-autoprovisioning-autoscale/1-install | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:privileged added: "deploy-elasticsearch" logger.go:42: 07:31:43 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 6] logger.go:42: 07:31:49 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_0.yml -n $NAMESPACE] logger.go:42: 07:31:50 | streaming-with-autoprovisioning-autoscale/1-install | statefulset.apps/elasticsearch created logger.go:42: 07:31:50 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 3] logger.go:42: 07:31:53 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_1.yml -n $NAMESPACE] logger.go:42: 07:31:54 | streaming-with-autoprovisioning-autoscale/1-install | service/elasticsearch created logger.go:42: 07:32:11 | streaming-with-autoprovisioning-autoscale/1-install | test step completed 1-install logger.go:42: 07:32:11 | streaming-with-autoprovisioning-autoscale/2-install | starting test step 2-install logger.go:42: 07:32:11 | streaming-with-autoprovisioning-autoscale/2-install | Jaeger:kuttl-test-grateful-goldfish/auto-provisioned created logger.go:42: 07:32:11 | streaming-with-autoprovisioning-autoscale/2-install | test step completed 2-install logger.go:42: 07:32:11 | streaming-with-autoprovisioning-autoscale/3- | starting test step 3- logger.go:42: 07:42:12 | streaming-with-autoprovisioning-autoscale/3- | test step failed 3- case.go:364: failed in step 3- case.go:366: --- StrimziPodSet:kuttl-test-grateful-goldfish/auto-provisioned-zookeeper +++ StrimziPodSet:kuttl-test-grateful-goldfish/auto-provisioned-zookeeper @@ -1,8 +1,180 @@ apiVersion: core.strimzi.io/v1beta2 kind: StrimziPodSet metadata: + annotations: + strimzi.io/storage: '{"type":"persistent-claim","size":"10Gi"}' + labels: + app: jaeger + app.kubernetes.io/instance: auto-provisioned + app.kubernetes.io/managed-by: strimzi-cluster-operator + app.kubernetes.io/name: zookeeper + app.kubernetes.io/part-of: strimzi-auto-provisioned + strimzi.io/cluster: auto-provisioned + strimzi.io/component-type: zookeeper + strimzi.io/kind: Kafka + strimzi.io/name: auto-provisioned-zookeeper + managedFields: + - apiVersion: core.strimzi.io/v1beta2 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + .: {} + f:strimzi.io/storage: {} + f:labels: + .: {} + f:app: {} + f:app.kubernetes.io/instance: {} + f:app.kubernetes.io/managed-by: {} + f:app.kubernetes.io/name: {} + f:app.kubernetes.io/part-of: {} + f:strimzi.io/cluster: {} + f:strimzi.io/component-type: {} + f:strimzi.io/kind: {} + f:strimzi.io/name: {} + f:ownerReferences: + .: {} + k:{"uid":"ab196284-dfb0-4720-8ae2-d4256b011c02"}: {} + f:spec: + .: {} + f:pods: {} + f:selector: + .: {} + f:matchLabels: + .: {} + f:strimzi.io/cluster: {} + f:strimzi.io/kind: {} + f:strimzi.io/name: {} + manager: strimzi-cluster-operator + operation: Update + time: "2024-09-26T07:32:17Z" name: auto-provisioned-zookeeper namespace: kuttl-test-grateful-goldfish -status: - readyPods: 1 + ownerReferences: + - apiVersion: kafka.strimzi.io/v1beta2 + blockOwnerDeletion: false + controller: false + kind: Kafka + name: auto-provisioned + uid: ab196284-dfb0-4720-8ae2-d4256b011c02 +spec: + pods: + - apiVersion: v1 + kind: Pod + metadata: + annotations: + strimzi.io/cluster-ca-cert-generation: "0" + strimzi.io/cluster-ca-key-generation: "0" + strimzi.io/logging-hash: 0f057cb0 + strimzi.io/revision: 91dfb950 + strimzi.io/server-cert-hash: dbdee6a9979b7d759a238728656cd67358d55d67 + labels: + app: jaeger + app.kubernetes.io/instance: auto-provisioned + app.kubernetes.io/managed-by: strimzi-cluster-operator + app.kubernetes.io/name: zookeeper + app.kubernetes.io/part-of: strimzi-auto-provisioned + com.company: Red_Hat + rht.comp: AMQ_Streams + rht.comp_ver: "2.7" + rht.prod_name: Red_Hat_Application_Foundations + rht.prod_ver: 2024.Q2 + rht.subcomp: zookeeper + rht.subcomp_t: infrastructure + statefulset.kubernetes.io/pod-name: auto-provisioned-zookeeper-0 + strimzi.io/cluster: auto-provisioned + strimzi.io/component-type: zookeeper + strimzi.io/controller: strimzipodset + strimzi.io/controller-name: auto-provisioned-zookeeper + strimzi.io/kind: Kafka + strimzi.io/name: auto-provisioned-zookeeper + strimzi.io/pod-name: auto-provisioned-zookeeper-0 + name: auto-provisioned-zookeeper-0 + namespace: kuttl-test-grateful-goldfish + spec: + containers: + - args: + - /opt/kafka/zookeeper_run.sh + env: + - name: ZOOKEEPER_METRICS_ENABLED + value: "false" + - name: ZOOKEEPER_SNAPSHOT_CHECK_ENABLED + value: "true" + - name: STRIMZI_KAFKA_GC_LOG_ENABLED + value: "false" + - name: KAFKA_HEAP_OPTS + value: -Xms128M + - name: ZOOKEEPER_CONFIGURATION + value: | + tickTime=2000 + initLimit=5 + syncLimit=2 + autopurge.purgeInterval=1 + admin.enableServer=false + image: registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:42bf60ce31540dd61fab2c9886d791e41f063ea6f86628694b9e60e49bc8951b + imagePullPolicy: IfNotPresent + livenessProbe: + exec: + command: + - /opt/kafka/zookeeper_healthcheck.sh + initialDelaySeconds: 15 + timeoutSeconds: 5 + name: zookeeper + ports: + - containerPort: 2888 + name: tcp-clustering + protocol: TCP + - containerPort: 3888 + name: tcp-election + protocol: TCP + - containerPort: 2181 + name: tcp-clients + protocol: TCP + readinessProbe: + exec: + command: + - /opt/kafka/zookeeper_healthcheck.sh + initialDelaySeconds: 15 + timeoutSeconds: 5 + volumeMounts: + - mountPath: /tmp + name: strimzi-tmp + - mountPath: /var/lib/zookeeper + name: data + - mountPath: /opt/kafka/custom-config/ + name: zookeeper-metrics-and-logging + - mountPath: /opt/kafka/zookeeper-node-certs/ + name: zookeeper-nodes + - mountPath: /opt/kafka/cluster-ca-certs/ + name: cluster-ca-certs + hostname: auto-provisioned-zookeeper-0 + restartPolicy: Always + schedulerName: default-scheduler + serviceAccountName: auto-provisioned-zookeeper + subdomain: auto-provisioned-zookeeper-nodes + terminationGracePeriodSeconds: 30 + volumes: + - emptyDir: + medium: Memory + sizeLimit: 5Mi + name: strimzi-tmp + - configMap: + name: auto-provisioned-zookeeper-config + name: zookeeper-metrics-and-logging + - name: zookeeper-nodes + secret: + defaultMode: 288 + secretName: auto-provisioned-zookeeper-nodes + - name: cluster-ca-certs + secret: + defaultMode: 288 + secretName: auto-provisioned-cluster-ca-cert + - name: data + persistentVolumeClaim: + claimName: data-auto-provisioned-zookeeper-0 + selector: + matchLabels: + strimzi.io/cluster: auto-provisioned + strimzi.io/kind: Kafka + strimzi.io/name: auto-provisioned-zookeeper case.go:366: resource StrimziPodSet:kuttl-test-grateful-goldfish/auto-provisioned-zookeeper: .status: key is missing from map logger.go:42: 07:42:12 | streaming-with-autoprovisioning-autoscale | streaming-with-autoprovisioning-autoscale events from ns kuttl-test-grateful-goldfish: logger.go:42: 07:42:12 | streaming-with-autoprovisioning-autoscale | 2024-09-26 07:31:50 +0000 UTC Normal Pod elasticsearch-0 Binding Scheduled Successfully assigned kuttl-test-grateful-goldfish/elasticsearch-0 to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:42:12 | streaming-with-autoprovisioning-autoscale | 2024-09-26 07:31:50 +0000 UTC Normal StatefulSet.apps elasticsearch SuccessfulCreate create Pod elasticsearch-0 in StatefulSet elasticsearch successful statefulset-controller logger.go:42: 07:42:12 | streaming-with-autoprovisioning-autoscale | 2024-09-26 07:31:51 +0000 UTC Normal Pod elasticsearch-0 AddedInterface Add eth0 [10.128.0.70/23] from ovn-kubernetes logger.go:42: 07:42:12 | streaming-with-autoprovisioning-autoscale | 2024-09-26 07:31:51 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulling Pulling image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" kubelet logger.go:42: 07:42:12 | streaming-with-autoprovisioning-autoscale | 2024-09-26 07:32:00 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulled Successfully pulled image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" in 9.573s (9.573s including waiting) kubelet logger.go:42: 07:42:12 | streaming-with-autoprovisioning-autoscale | 2024-09-26 07:32:00 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:42:12 | streaming-with-autoprovisioning-autoscale | 2024-09-26 07:32:00 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:42:12 | streaming-with-autoprovisioning-autoscale | 2024-09-26 07:32:05 +0000 UTC Warning Pod elasticsearch-0.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Get "http://10.128.0.70:9200/": dial tcp 10.128.0.70:9200: connect: connection refused kubelet logger.go:42: 07:42:12 | streaming-with-autoprovisioning-autoscale | 2024-09-26 07:32:16 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 07:42:12 | streaming-with-autoprovisioning-autoscale | 2024-09-26 07:32:17 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-zookeeper NoPods No matching pods found controllermanager logger.go:42: 07:42:12 | streaming-with-autoprovisioning-autoscale | Deleting namespace: kuttl-test-grateful-goldfish === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- FAIL: kuttl (643.39s) --- FAIL: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.02s) --- FAIL: kuttl/harness/streaming-with-autoprovisioning-autoscale (636.75s) FAIL + exit_code=1 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name streaming --report --output /logs/artifacts/streaming.xml ./artifacts/kuttl-report.xml time="2024-09-26T07:42:26Z" level=debug msg="Setting a new name for the test suites" time="2024-09-26T07:42:26Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-09-26T07:42:26Z" level=debug msg="normalizing test case names" time="2024-09-26T07:42:26Z" level=debug msg="streaming/artifacts -> streaming_artifacts" time="2024-09-26T07:42:26Z" level=debug msg="streaming/streaming-with-autoprovisioning-autoscale -> streaming_streaming_with_autoprovisioning_autoscale" +-----------------------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------------------+--------+ | streaming_artifacts | passed | | streaming_streaming_with_autoprovisioning_autoscale | failed | +-----------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh ui false true + '[' 3 -ne 3 ']' + test_suite_name=ui + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/ui.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-ui make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true ./tests/e2e/ui/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.32 True False 55m Cluster version is 4.15.32' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.32 True False 55m Cluster version is 4.15.32' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/ui/render.sh ++ export SUITE_DIR=./tests/e2e/ui ++ SUITE_DIR=./tests/e2e/ui ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/ui ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test allinone + '[' 1 -ne 1 ']' + test_name=allinone + echo =========================================================================== =========================================================================== + info 'Rendering files for test allinone' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test allinone\e[0m' Rendering files for test allinone + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build + '[' _build '!=' _build ']' + mkdir -p allinone + cd allinone + export GET_URL_COMMAND + export URL + export JAEGER_NAME=all-in-one-ui + JAEGER_NAME=all-in-one-ui + '[' true = true ']' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./01-curl.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./04-test-ui-config.yaml + start_test production + '[' 1 -ne 1 ']' + test_name=production + echo =========================================================================== =========================================================================== + info 'Rendering files for test production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test production\e[0m' Rendering files for test production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build/allinone + '[' allinone '!=' _build ']' + cd .. + mkdir -p production + cd production + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + [[ true = true ]] + [[ true = true ]] + render_install_jaeger production-ui production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + '[' true = true ']' + INSECURE=true + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-forbbiden-access.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-curl.yaml + INSECURE=true + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./05-check-disabled-security.yaml + ASSERT_PRESENT=false + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./06-check-NO-gaID.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./08-check-gaID.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running ui E2E tests' Running ui E2E tests + cd tests/e2e/ui/_build + set +e + KUBECONFIG=/tmp/kubeconfig-488946273 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-rosa-h-ezi9.l97m.s3.devshift.org:443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 3 tests === RUN kuttl/harness === RUN kuttl/harness/allinone === PAUSE kuttl/harness/allinone === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/production === PAUSE kuttl/harness/production === CONT kuttl/harness/allinone logger.go:42: 07:42:34 | allinone | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:42:34 | allinone | Creating namespace: kuttl-test-driven-anchovy logger.go:42: 07:42:34 | allinone/0-install | starting test step 0-install logger.go:42: 07:42:34 | allinone/0-install | Jaeger:kuttl-test-driven-anchovy/all-in-one-ui created logger.go:42: 07:42:35 | allinone/0-install | test step completed 0-install logger.go:42: 07:42:35 | allinone/1-curl | starting test step 1-curl logger.go:42: 07:42:35 | allinone/1-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 07:42:35 | allinone/1-curl | Checking the Ingress host value was populated logger.go:42: 07:42:35 | allinone/1-curl | Try number 0 logger.go:42: 07:42:36 | allinone/1-curl | error: error executing jsonpath "{.items[0].status.ingress[0].host}": Error executing template: array index out of bounds: index 0, length 0. Printing more information for debugging the template: logger.go:42: 07:42:36 | allinone/1-curl | template was: logger.go:42: 07:42:36 | allinone/1-curl | {.items[0].status.ingress[0].host} logger.go:42: 07:42:36 | allinone/1-curl | object given to jsonpath engine was: logger.go:42: 07:42:36 | allinone/1-curl | map[string]interface {}{"apiVersion":"v1", "items":[]interface {}{}, "kind":"List", "metadata":map[string]interface {}{"resourceVersion":""}} logger.go:42: 07:42:36 | allinone/1-curl | logger.go:42: 07:42:36 | allinone/1-curl | logger.go:42: 07:42:46 | allinone/1-curl | Try number 1 logger.go:42: 07:42:46 | allinone/1-curl | Hostname is all-in-one-ui-kuttl-test-driven-anchovy.apps.rosa.ci-rosa-h-ezi9.l97m.s3.devshift.org logger.go:42: 07:42:46 | allinone/1-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE all-in-one-ui] logger.go:42: 07:42:46 | allinone/1-curl | Checking an expected HTTP response logger.go:42: 07:42:46 | allinone/1-curl | Running in OpenShift logger.go:42: 07:42:46 | allinone/1-curl | User not provided. Getting the token... logger.go:42: 07:42:48 | allinone/1-curl | Warning: resource jaegers/all-in-one-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:42:55 | allinone/1-curl | Try number 1/30 the https://all-in-one-ui-kuttl-test-driven-anchovy.apps.rosa.ci-rosa-h-ezi9.l97m.s3.devshift.org/search logger.go:42: 07:42:55 | allinone/1-curl | curl response asserted properly logger.go:42: 07:42:56 | allinone/1-curl | test step completed 1-curl logger.go:42: 07:42:56 | allinone/2-delete | starting test step 2-delete logger.go:42: 07:42:56 | allinone/2-delete | Jaeger:kuttl-test-driven-anchovy/all-in-one-ui created logger.go:42: 07:42:56 | allinone/2-delete | test step completed 2-delete logger.go:42: 07:42:56 | allinone/3-install | starting test step 3-install logger.go:42: 07:42:56 | allinone/3-install | Jaeger:kuttl-test-driven-anchovy/all-in-one-ui updated logger.go:42: 07:42:56 | allinone/3-install | test step completed 3-install logger.go:42: 07:42:56 | allinone/4-test-ui-config | starting test step 4-test-ui-config logger.go:42: 07:42:56 | allinone/4-test-ui-config | running command: [./ensure-ingress-host.sh] logger.go:42: 07:42:56 | allinone/4-test-ui-config | Checking the Ingress host value was populated logger.go:42: 07:42:56 | allinone/4-test-ui-config | Try number 0 logger.go:42: 07:42:57 | allinone/4-test-ui-config | error: error executing jsonpath "{.items[0].status.ingress[0].host}": Error executing template: array index out of bounds: index 0, length 0. Printing more information for debugging the template: logger.go:42: 07:42:57 | allinone/4-test-ui-config | template was: logger.go:42: 07:42:57 | allinone/4-test-ui-config | {.items[0].status.ingress[0].host} logger.go:42: 07:42:57 | allinone/4-test-ui-config | object given to jsonpath engine was: logger.go:42: 07:42:57 | allinone/4-test-ui-config | map[string]interface {}{"apiVersion":"v1", "items":[]interface {}{}, "kind":"List", "metadata":map[string]interface {}{"resourceVersion":""}} logger.go:42: 07:42:57 | allinone/4-test-ui-config | logger.go:42: 07:42:57 | allinone/4-test-ui-config | logger.go:42: 07:43:07 | allinone/4-test-ui-config | Try number 1 logger.go:42: 07:43:07 | allinone/4-test-ui-config | Hostname is all-in-one-ui-kuttl-test-driven-anchovy.apps.rosa.ci-rosa-h-ezi9.l97m.s3.devshift.org logger.go:42: 07:43:07 | allinone/4-test-ui-config | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 07:43:08 | allinone/4-test-ui-config | time="2024-09-26T07:43:08Z" level=info msg="Querying https://all-in-one-ui-kuttl-test-driven-anchovy.apps.rosa.ci-rosa-h-ezi9.l97m.s3.devshift.org/search..." logger.go:42: 07:43:08 | allinone/4-test-ui-config | time="2024-09-26T07:43:08Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 07:43:08 | allinone/4-test-ui-config | time="2024-09-26T07:43:08Z" level=info msg="Polling to https://all-in-one-ui-kuttl-test-driven-anchovy.apps.rosa.ci-rosa-h-ezi9.l97m.s3.devshift.org/search" logger.go:42: 07:43:08 | allinone/4-test-ui-config | time="2024-09-26T07:43:08Z" level=info msg="Doing request number 0" logger.go:42: 07:43:08 | allinone/4-test-ui-config | time="2024-09-26T07:43:08Z" level=info msg="Content found and asserted!" logger.go:42: 07:43:08 | allinone/4-test-ui-config | time="2024-09-26T07:43:08Z" level=info msg="Success!" logger.go:42: 07:43:08 | allinone/4-test-ui-config | test step completed 4-test-ui-config logger.go:42: 07:43:08 | allinone | allinone events from ns kuttl-test-driven-anchovy: logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:35 +0000 UTC Normal Pod all-in-one-ui-7fc5f58bd9-9tkl5 Binding Scheduled Successfully assigned kuttl-test-driven-anchovy/all-in-one-ui-7fc5f58bd9-9tkl5 to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:35 +0000 UTC Normal Pod all-in-one-ui-7fc5f58bd9-9tkl5 AddedInterface Add eth0 [10.128.0.71/23] from ovn-kubernetes logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:35 +0000 UTC Normal Pod all-in-one-ui-7fc5f58bd9-9tkl5.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:35 +0000 UTC Normal Pod all-in-one-ui-7fc5f58bd9-9tkl5.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:35 +0000 UTC Normal Pod all-in-one-ui-7fc5f58bd9-9tkl5.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:35 +0000 UTC Normal Pod all-in-one-ui-7fc5f58bd9-9tkl5.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:35 +0000 UTC Normal Pod all-in-one-ui-7fc5f58bd9-9tkl5.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:35 +0000 UTC Normal Pod all-in-one-ui-7fc5f58bd9-9tkl5.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:35 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-7fc5f58bd9 SuccessfulCreate Created pod: all-in-one-ui-7fc5f58bd9-9tkl5 replicaset-controller logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:35 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-7fc5f58bd9 to 1 deployment-controller logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:49 +0000 UTC Normal Pod all-in-one-ui-7fc5f58bd9-9tkl5.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:49 +0000 UTC Normal Pod all-in-one-ui-7fc5f58bd9-9tkl5.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:49 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-7fc5f58bd9 SuccessfulDelete Deleted pod: all-in-one-ui-7fc5f58bd9-9tkl5 replicaset-controller logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:49 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled down replica set all-in-one-ui-7fc5f58bd9 to 0 from 1 deployment-controller logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:50 +0000 UTC Normal Pod all-in-one-ui-74fdbb788c-vcwb4 Binding Scheduled Successfully assigned kuttl-test-driven-anchovy/all-in-one-ui-74fdbb788c-vcwb4 to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:50 +0000 UTC Normal Pod all-in-one-ui-74fdbb788c-vcwb4 AddedInterface Add eth0 [10.129.0.90/23] from ovn-kubernetes logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:50 +0000 UTC Normal Pod all-in-one-ui-74fdbb788c-vcwb4.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:50 +0000 UTC Normal Pod all-in-one-ui-74fdbb788c-vcwb4.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:50 +0000 UTC Normal Pod all-in-one-ui-74fdbb788c-vcwb4.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:50 +0000 UTC Normal Pod all-in-one-ui-74fdbb788c-vcwb4.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:50 +0000 UTC Normal Pod all-in-one-ui-74fdbb788c-vcwb4.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:50 +0000 UTC Normal Pod all-in-one-ui-74fdbb788c-vcwb4.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:50 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-74fdbb788c SuccessfulCreate Created pod: all-in-one-ui-74fdbb788c-vcwb4 replicaset-controller logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:50 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-74fdbb788c to 1 deployment-controller logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:56 +0000 UTC Normal Pod all-in-one-ui-74fdbb788c-vcwb4.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:56 +0000 UTC Normal Pod all-in-one-ui-74fdbb788c-vcwb4.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:57 +0000 UTC Normal Pod all-in-one-ui-7c6b65d56d-49jhk Binding Scheduled Successfully assigned kuttl-test-driven-anchovy/all-in-one-ui-7c6b65d56d-49jhk to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:57 +0000 UTC Normal Pod all-in-one-ui-7c6b65d56d-49jhk AddedInterface Add eth0 [10.128.0.72/23] from ovn-kubernetes logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:57 +0000 UTC Normal Pod all-in-one-ui-7c6b65d56d-49jhk.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:57 +0000 UTC Normal Pod all-in-one-ui-7c6b65d56d-49jhk.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:57 +0000 UTC Normal Pod all-in-one-ui-7c6b65d56d-49jhk.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:57 +0000 UTC Normal Pod all-in-one-ui-7c6b65d56d-49jhk.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:57 +0000 UTC Normal Pod all-in-one-ui-7c6b65d56d-49jhk.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:57 +0000 UTC Normal Pod all-in-one-ui-7c6b65d56d-49jhk.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:57 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-7c6b65d56d SuccessfulCreate Created pod: all-in-one-ui-7c6b65d56d-49jhk replicaset-controller logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:57 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-7c6b65d56d to 1 deployment-controller logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:58 +0000 UTC Warning Pod all-in-one-ui-7c6b65d56d-49jhk FailedToRetrieveImagePullSecret Unable to retrieve some image pull secrets (all-in-one-ui-ui-proxy-dockercfg-mqbj9); attempting to pull the image may not succeed. kubelet logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:58 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-7c6b65d56d SuccessfulDelete Deleted pod: all-in-one-ui-7c6b65d56d-49jhk replicaset-controller logger.go:42: 07:43:08 | allinone | 2024-09-26 07:42:58 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled down replica set all-in-one-ui-7c6b65d56d to 0 from 1 deployment-controller logger.go:42: 07:43:08 | allinone | 2024-09-26 07:43:00 +0000 UTC Normal Pod all-in-one-ui-7c6b65d56d-49jhk.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:43:08 | allinone | 2024-09-26 07:43:00 +0000 UTC Normal Pod all-in-one-ui-7c6b65d56d-49jhk.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:43:08 | allinone | 2024-09-26 07:43:01 +0000 UTC Normal Pod all-in-one-ui-79bb466bb4-kmnss Binding Scheduled Successfully assigned kuttl-test-driven-anchovy/all-in-one-ui-79bb466bb4-kmnss to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:43:08 | allinone | 2024-09-26 07:43:01 +0000 UTC Normal Pod all-in-one-ui-79bb466bb4-kmnss AddedInterface Add eth0 [10.128.0.73/23] from ovn-kubernetes logger.go:42: 07:43:08 | allinone | 2024-09-26 07:43:01 +0000 UTC Normal Pod all-in-one-ui-79bb466bb4-kmnss.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:43:08 | allinone | 2024-09-26 07:43:01 +0000 UTC Normal Pod all-in-one-ui-79bb466bb4-kmnss.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:43:08 | allinone | 2024-09-26 07:43:01 +0000 UTC Normal Pod all-in-one-ui-79bb466bb4-kmnss.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:43:08 | allinone | 2024-09-26 07:43:01 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-79bb466bb4 SuccessfulCreate Created pod: all-in-one-ui-79bb466bb4-kmnss replicaset-controller logger.go:42: 07:43:08 | allinone | 2024-09-26 07:43:01 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-79bb466bb4 to 1 deployment-controller logger.go:42: 07:43:08 | allinone | Deleting namespace: kuttl-test-driven-anchovy === CONT kuttl/harness/production logger.go:42: 07:43:15 | production | Ignoring add-tracking-id.yaml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:43:15 | production | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:43:15 | production | Creating namespace: kuttl-test-tender-ghoul logger.go:42: 07:43:15 | production/1-install | starting test step 1-install logger.go:42: 07:43:15 | production/1-install | Jaeger:kuttl-test-tender-ghoul/production-ui created logger.go:42: 07:43:50 | production/1-install | test step completed 1-install logger.go:42: 07:43:50 | production/2-check-forbbiden-access | starting test step 2-check-forbbiden-access logger.go:42: 07:43:50 | production/2-check-forbbiden-access | running command: [./ensure-ingress-host.sh] logger.go:42: 07:43:50 | production/2-check-forbbiden-access | Checking the Ingress host value was populated logger.go:42: 07:43:50 | production/2-check-forbbiden-access | Try number 0 logger.go:42: 07:43:50 | production/2-check-forbbiden-access | error: error executing jsonpath "{.items[0].status.ingress[0].host}": Error executing template: array index out of bounds: index 0, length 0. Printing more information for debugging the template: logger.go:42: 07:43:50 | production/2-check-forbbiden-access | template was: logger.go:42: 07:43:50 | production/2-check-forbbiden-access | {.items[0].status.ingress[0].host} logger.go:42: 07:43:50 | production/2-check-forbbiden-access | object given to jsonpath engine was: logger.go:42: 07:43:50 | production/2-check-forbbiden-access | map[string]interface {}{"apiVersion":"v1", "items":[]interface {}{}, "kind":"List", "metadata":map[string]interface {}{"resourceVersion":""}} logger.go:42: 07:43:50 | production/2-check-forbbiden-access | logger.go:42: 07:43:50 | production/2-check-forbbiden-access | logger.go:42: 07:44:00 | production/2-check-forbbiden-access | Try number 1 logger.go:42: 07:44:00 | production/2-check-forbbiden-access | Hostname is production-ui-kuttl-test-tender-ghoul.apps.rosa.ci-rosa-h-ezi9.l97m.s3.devshift.org logger.go:42: 07:44:00 | production/2-check-forbbiden-access | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE production-ui] logger.go:42: 07:44:00 | production/2-check-forbbiden-access | Checking an expected HTTP response logger.go:42: 07:44:00 | production/2-check-forbbiden-access | Running in OpenShift logger.go:42: 07:44:00 | production/2-check-forbbiden-access | Not using any secret logger.go:42: 07:44:00 | production/2-check-forbbiden-access | Try number 1/30 the https://production-ui-kuttl-test-tender-ghoul.apps.rosa.ci-rosa-h-ezi9.l97m.s3.devshift.org/search logger.go:42: 07:44:01 | production/2-check-forbbiden-access | curl response asserted properly logger.go:42: 07:44:01 | production/2-check-forbbiden-access | test step completed 2-check-forbbiden-access logger.go:42: 07:44:01 | production/3-curl | starting test step 3-curl logger.go:42: 07:44:01 | production/3-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 07:44:01 | production/3-curl | Checking the Ingress host value was populated logger.go:42: 07:44:01 | production/3-curl | Try number 0 logger.go:42: 07:44:01 | production/3-curl | Hostname is production-ui-kuttl-test-tender-ghoul.apps.rosa.ci-rosa-h-ezi9.l97m.s3.devshift.org logger.go:42: 07:44:01 | production/3-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 07:44:01 | production/3-curl | Checking an expected HTTP response logger.go:42: 07:44:01 | production/3-curl | Running in OpenShift logger.go:42: 07:44:01 | production/3-curl | User not provided. Getting the token... logger.go:42: 07:44:03 | production/3-curl | Warning: resource jaegers/production-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:44:10 | production/3-curl | Try number 1/30 the https://production-ui-kuttl-test-tender-ghoul.apps.rosa.ci-rosa-h-ezi9.l97m.s3.devshift.org/search logger.go:42: 07:44:11 | production/3-curl | curl response asserted properly logger.go:42: 07:44:11 | production/3-curl | test step completed 3-curl logger.go:42: 07:44:11 | production/4-install | starting test step 4-install logger.go:42: 07:44:11 | production/4-install | Jaeger:kuttl-test-tender-ghoul/production-ui updated logger.go:42: 07:44:11 | production/4-install | test step completed 4-install logger.go:42: 07:44:11 | production/5-check-disabled-security | starting test step 5-check-disabled-security logger.go:42: 07:44:11 | production/5-check-disabled-security | running command: [./ensure-ingress-host.sh] logger.go:42: 07:44:11 | production/5-check-disabled-security | Checking the Ingress host value was populated logger.go:42: 07:44:11 | production/5-check-disabled-security | Try number 0 logger.go:42: 07:44:11 | production/5-check-disabled-security | Hostname is production-ui-kuttl-test-tender-ghoul.apps.rosa.ci-rosa-h-ezi9.l97m.s3.devshift.org logger.go:42: 07:44:11 | production/5-check-disabled-security | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 07:44:12 | production/5-check-disabled-security | Checking an expected HTTP response logger.go:42: 07:44:12 | production/5-check-disabled-security | Running in OpenShift logger.go:42: 07:44:12 | production/5-check-disabled-security | Not using any secret logger.go:42: 07:44:12 | production/5-check-disabled-security | Try number 1/30 the https://production-ui-kuttl-test-tender-ghoul.apps.rosa.ci-rosa-h-ezi9.l97m.s3.devshift.org/search logger.go:42: 07:44:28 | production/5-check-disabled-security | HTTP response is 503. 200 expected. Waiting 10 s logger.go:42: 07:44:38 | production/5-check-disabled-security | Try number 2/30 the https://production-ui-kuttl-test-tender-ghoul.apps.rosa.ci-rosa-h-ezi9.l97m.s3.devshift.org/search logger.go:42: 07:44:38 | production/5-check-disabled-security | curl response asserted properly logger.go:42: 07:44:38 | production/5-check-disabled-security | test step completed 5-check-disabled-security logger.go:42: 07:44:38 | production/6-check-NO-gaID | starting test step 6-check-NO-gaID logger.go:42: 07:44:38 | production/6-check-NO-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 07:44:38 | production/6-check-NO-gaID | Checking the Ingress host value was populated logger.go:42: 07:44:38 | production/6-check-NO-gaID | Try number 0 logger.go:42: 07:44:39 | production/6-check-NO-gaID | Hostname is production-ui-kuttl-test-tender-ghoul.apps.rosa.ci-rosa-h-ezi9.l97m.s3.devshift.org logger.go:42: 07:44:39 | production/6-check-NO-gaID | running command: [sh -c ASSERT_PRESENT=false EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 07:44:39 | production/6-check-NO-gaID | time="2024-09-26T07:44:39Z" level=info msg="Querying https://production-ui-kuttl-test-tender-ghoul.apps.rosa.ci-rosa-h-ezi9.l97m.s3.devshift.org/search..." logger.go:42: 07:44:39 | production/6-check-NO-gaID | time="2024-09-26T07:44:39Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 07:44:39 | production/6-check-NO-gaID | time="2024-09-26T07:44:39Z" level=info msg="Polling to https://production-ui-kuttl-test-tender-ghoul.apps.rosa.ci-rosa-h-ezi9.l97m.s3.devshift.org/search" logger.go:42: 07:44:39 | production/6-check-NO-gaID | time="2024-09-26T07:44:39Z" level=info msg="Doing request number 0" logger.go:42: 07:44:39 | production/6-check-NO-gaID | time="2024-09-26T07:44:39Z" level=info msg="Content not found and asserted it was not found!" logger.go:42: 07:44:39 | production/6-check-NO-gaID | time="2024-09-26T07:44:39Z" level=info msg="Success!" logger.go:42: 07:44:39 | production/6-check-NO-gaID | test step completed 6-check-NO-gaID logger.go:42: 07:44:39 | production/7-add-tracking-id | starting test step 7-add-tracking-id logger.go:42: 07:44:39 | production/7-add-tracking-id | running command: [sh -c kubectl apply -f add-tracking-id.yaml -n $NAMESPACE] logger.go:42: 07:44:40 | production/7-add-tracking-id | jaeger.jaegertracing.io/production-ui configured logger.go:42: 07:44:40 | production/7-add-tracking-id | test step completed 7-add-tracking-id logger.go:42: 07:44:40 | production/8-check-gaID | starting test step 8-check-gaID logger.go:42: 07:44:40 | production/8-check-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 07:44:40 | production/8-check-gaID | Checking the Ingress host value was populated logger.go:42: 07:44:40 | production/8-check-gaID | Try number 0 logger.go:42: 07:44:41 | production/8-check-gaID | Hostname is production-ui-kuttl-test-tender-ghoul.apps.rosa.ci-rosa-h-ezi9.l97m.s3.devshift.org logger.go:42: 07:44:41 | production/8-check-gaID | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 07:44:41 | production/8-check-gaID | time="2024-09-26T07:44:41Z" level=info msg="Querying https://production-ui-kuttl-test-tender-ghoul.apps.rosa.ci-rosa-h-ezi9.l97m.s3.devshift.org/search..." logger.go:42: 07:44:41 | production/8-check-gaID | time="2024-09-26T07:44:41Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 07:44:41 | production/8-check-gaID | time="2024-09-26T07:44:41Z" level=info msg="Polling to https://production-ui-kuttl-test-tender-ghoul.apps.rosa.ci-rosa-h-ezi9.l97m.s3.devshift.org/search" logger.go:42: 07:44:41 | production/8-check-gaID | time="2024-09-26T07:44:41Z" level=info msg="Doing request number 0" logger.go:42: 07:44:41 | production/8-check-gaID | time="2024-09-26T07:44:41Z" level=warning msg="Status code: 503" logger.go:42: 07:44:49 | production/8-check-gaID | time="2024-09-26T07:44:49Z" level=info msg="Doing request number 1" logger.go:42: 07:44:49 | production/8-check-gaID | time="2024-09-26T07:44:49Z" level=info msg="Content found and asserted!" logger.go:42: 07:44:49 | production/8-check-gaID | time="2024-09-26T07:44:49Z" level=info msg="Success!" logger.go:42: 07:44:49 | production/8-check-gaID | test step completed 8-check-gaID logger.go:42: 07:44:50 | production | production events from ns kuttl-test-tender-ghoul: logger.go:42: 07:44:50 | production | 2024-09-26 07:43:19 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltesttenderghoulproductionui-1-7c54fdbf87 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltesttenderghoulproductionui-1-7c54fg6qg5 replicaset-controller logger.go:42: 07:44:50 | production | 2024-09-26 07:43:19 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttenderghoulproductionui-1-7c54fg6qg5 Binding Scheduled Successfully assigned kuttl-test-tender-ghoul/elasticsearch-cdm-kuttltesttenderghoulproductionui-1-7c54fg6qg5 to ip-10-0-61-187.us-west-2.compute.internal default-scheduler logger.go:42: 07:44:50 | production | 2024-09-26 07:43:19 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltesttenderghoulproductionui-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltesttenderghoulproductionui-1-7c54fdbf87 to 1 deployment-controller logger.go:42: 07:44:50 | production | 2024-09-26 07:43:20 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttenderghoulproductionui-1-7c54fg6qg5 AddedInterface Add eth0 [10.128.0.74/23] from ovn-kubernetes logger.go:42: 07:44:50 | production | 2024-09-26 07:43:20 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttenderghoulproductionui-1-7c54fg6qg5.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:bc97b8e13087050dfb1a75b02c5b780cbb8fb12a849a655169d072ba8bbf42b4" already present on machine kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:43:20 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttenderghoulproductionui-1-7c54fg6qg5.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:43:20 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttenderghoulproductionui-1-7c54fg6qg5.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:43:20 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttenderghoulproductionui-1-7c54fg6qg5.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:1d1d7d47b616995d18692f3a2d6232d7e1a0f41bc1503d87100f31939e080a78" already present on machine kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:43:20 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttenderghoulproductionui-1-7c54fg6qg5.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:43:20 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesttenderghoulproductionui-1-7c54fg6qg5.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:43:35 +0000 UTC Warning Pod elasticsearch-cdm-kuttltesttenderghoulproductionui-1-7c54fg6qg5.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:43:46 +0000 UTC Normal Pod production-ui-collector-7f5cdcbd6-mrqfc Binding Scheduled Successfully assigned kuttl-test-tender-ghoul/production-ui-collector-7f5cdcbd6-mrqfc to ip-10-0-51-77.us-west-2.compute.internal default-scheduler logger.go:42: 07:44:50 | production | 2024-09-26 07:43:46 +0000 UTC Warning Pod production-ui-collector-7f5cdcbd6-mrqfc FailedMount MountVolume.SetUp failed for volume "production-ui-collector-tls-config-volume" : secret "production-ui-collector-headless-tls" not found kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:43:46 +0000 UTC Normal ReplicaSet.apps production-ui-collector-7f5cdcbd6 SuccessfulCreate Created pod: production-ui-collector-7f5cdcbd6-mrqfc replicaset-controller logger.go:42: 07:44:50 | production | 2024-09-26 07:43:46 +0000 UTC Normal Deployment.apps production-ui-collector ScalingReplicaSet Scaled up replica set production-ui-collector-7f5cdcbd6 to 1 deployment-controller logger.go:42: 07:44:50 | production | 2024-09-26 07:43:46 +0000 UTC Normal Pod production-ui-query-69bf99779c-txfw7 Binding Scheduled Successfully assigned kuttl-test-tender-ghoul/production-ui-query-69bf99779c-txfw7 to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:44:50 | production | 2024-09-26 07:43:46 +0000 UTC Warning Pod production-ui-query-69bf99779c-txfw7 FailedMount MountVolume.SetUp failed for volume "production-ui-ui-oauth-proxy-tls" : secret "production-ui-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:43:46 +0000 UTC Normal ReplicaSet.apps production-ui-query-69bf99779c SuccessfulCreate Created pod: production-ui-query-69bf99779c-txfw7 replicaset-controller logger.go:42: 07:44:50 | production | 2024-09-26 07:43:46 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-69bf99779c to 1 deployment-controller logger.go:42: 07:44:50 | production | 2024-09-26 07:43:47 +0000 UTC Normal Pod production-ui-collector-7f5cdcbd6-mrqfc AddedInterface Add eth0 [10.129.0.91/23] from ovn-kubernetes logger.go:42: 07:44:50 | production | 2024-09-26 07:43:47 +0000 UTC Normal Pod production-ui-collector-7f5cdcbd6-mrqfc.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:43:47 +0000 UTC Normal Pod production-ui-collector-7f5cdcbd6-mrqfc.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:43:47 +0000 UTC Normal Pod production-ui-collector-7f5cdcbd6-mrqfc.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:43:47 +0000 UTC Normal Pod production-ui-query-69bf99779c-txfw7 AddedInterface Add eth0 [10.130.0.77/23] from ovn-kubernetes logger.go:42: 07:44:50 | production | 2024-09-26 07:43:47 +0000 UTC Normal Pod production-ui-query-69bf99779c-txfw7.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:43:47 +0000 UTC Normal Pod production-ui-query-69bf99779c-txfw7.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:43:47 +0000 UTC Normal Pod production-ui-query-69bf99779c-txfw7.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:43:47 +0000 UTC Normal Pod production-ui-query-69bf99779c-txfw7.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:43:47 +0000 UTC Normal Pod production-ui-query-69bf99779c-txfw7.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:43:47 +0000 UTC Normal Pod production-ui-query-69bf99779c-txfw7.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:43:47 +0000 UTC Normal Pod production-ui-query-69bf99779c-txfw7.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:43:47 +0000 UTC Normal Pod production-ui-query-69bf99779c-txfw7.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:43:47 +0000 UTC Normal Pod production-ui-query-69bf99779c-txfw7.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:44:04 +0000 UTC Normal Pod production-ui-query-69bf99779c-txfw7.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:44:04 +0000 UTC Normal Pod production-ui-query-69bf99779c-txfw7.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:44:04 +0000 UTC Normal Pod production-ui-query-69bf99779c-txfw7.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:44:04 +0000 UTC Normal ReplicaSet.apps production-ui-query-69bf99779c SuccessfulDelete Deleted pod: production-ui-query-69bf99779c-txfw7 replicaset-controller logger.go:42: 07:44:50 | production | 2024-09-26 07:44:04 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-69bf99779c to 0 from 1 deployment-controller logger.go:42: 07:44:50 | production | 2024-09-26 07:44:05 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:44:50 | production | 2024-09-26 07:44:05 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:44:50 | production | 2024-09-26 07:44:05 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:44:50 | production | 2024-09-26 07:44:05 +0000 UTC Normal Pod production-ui-query-66c4d58659-7f4x4 Binding Scheduled Successfully assigned kuttl-test-tender-ghoul/production-ui-query-66c4d58659-7f4x4 to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:44:50 | production | 2024-09-26 07:44:05 +0000 UTC Normal ReplicaSet.apps production-ui-query-66c4d58659 SuccessfulCreate Created pod: production-ui-query-66c4d58659-7f4x4 replicaset-controller logger.go:42: 07:44:50 | production | 2024-09-26 07:44:05 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-66c4d58659 to 1 deployment-controller logger.go:42: 07:44:50 | production | 2024-09-26 07:44:06 +0000 UTC Normal Pod production-ui-query-66c4d58659-7f4x4 AddedInterface Add eth0 [10.130.0.78/23] from ovn-kubernetes logger.go:42: 07:44:50 | production | 2024-09-26 07:44:06 +0000 UTC Normal Pod production-ui-query-66c4d58659-7f4x4.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:44:06 +0000 UTC Normal Pod production-ui-query-66c4d58659-7f4x4.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:44:06 +0000 UTC Normal Pod production-ui-query-66c4d58659-7f4x4.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:44:06 +0000 UTC Normal Pod production-ui-query-66c4d58659-7f4x4.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:44:06 +0000 UTC Normal Pod production-ui-query-66c4d58659-7f4x4.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:44:06 +0000 UTC Normal Pod production-ui-query-66c4d58659-7f4x4.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:44:06 +0000 UTC Normal Pod production-ui-query-66c4d58659-7f4x4.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:44:06 +0000 UTC Normal Pod production-ui-query-66c4d58659-7f4x4.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:44:06 +0000 UTC Normal Pod production-ui-query-66c4d58659-7f4x4.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:44:12 +0000 UTC Normal Pod production-ui-query-66c4d58659-7f4x4.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:44:12 +0000 UTC Normal Pod production-ui-query-66c4d58659-7f4x4.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:44:12 +0000 UTC Normal Pod production-ui-query-66c4d58659-7f4x4.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:44:12 +0000 UTC Normal ReplicaSet.apps production-ui-query-66c4d58659 SuccessfulDelete Deleted pod: production-ui-query-66c4d58659-7f4x4 replicaset-controller logger.go:42: 07:44:50 | production | 2024-09-26 07:44:12 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-66c4d58659 to 0 from 1 deployment-controller logger.go:42: 07:44:50 | production | 2024-09-26 07:44:12 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-86d95f4fd7 to 1 deployment-controller logger.go:42: 07:44:50 | production | 2024-09-26 07:44:13 +0000 UTC Normal Pod production-ui-query-86d95f4fd7-8jqn4 Binding Scheduled Successfully assigned kuttl-test-tender-ghoul/production-ui-query-86d95f4fd7-8jqn4 to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:44:50 | production | 2024-09-26 07:44:13 +0000 UTC Normal Pod production-ui-query-86d95f4fd7-8jqn4 AddedInterface Add eth0 [10.130.0.79/23] from ovn-kubernetes logger.go:42: 07:44:50 | production | 2024-09-26 07:44:13 +0000 UTC Normal Pod production-ui-query-86d95f4fd7-8jqn4.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:44:13 +0000 UTC Normal Pod production-ui-query-86d95f4fd7-8jqn4.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:44:13 +0000 UTC Normal Pod production-ui-query-86d95f4fd7-8jqn4.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:44:13 +0000 UTC Normal Pod production-ui-query-86d95f4fd7-8jqn4.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:44:13 +0000 UTC Normal Pod production-ui-query-86d95f4fd7-8jqn4.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:44:13 +0000 UTC Normal Pod production-ui-query-86d95f4fd7-8jqn4.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:44:13 +0000 UTC Normal ReplicaSet.apps production-ui-query-86d95f4fd7 SuccessfulCreate Created pod: production-ui-query-86d95f4fd7-8jqn4 replicaset-controller logger.go:42: 07:44:50 | production | 2024-09-26 07:44:35 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod production-ui-collector-7f5cdcbd6-mrqfc horizontal-pod-autoscaler logger.go:42: 07:44:50 | production | 2024-09-26 07:44:41 +0000 UTC Normal Pod production-ui-query-86d95f4fd7-8jqn4.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:44:41 +0000 UTC Normal Pod production-ui-query-86d95f4fd7-8jqn4.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:44:41 +0000 UTC Normal ReplicaSet.apps production-ui-query-86d95f4fd7 SuccessfulDelete Deleted pod: production-ui-query-86d95f4fd7-8jqn4 replicaset-controller logger.go:42: 07:44:50 | production | 2024-09-26 07:44:41 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-86d95f4fd7 to 0 from 1 deployment-controller logger.go:42: 07:44:50 | production | 2024-09-26 07:44:42 +0000 UTC Normal Pod production-ui-query-6699c7c8cb-2mj2k Binding Scheduled Successfully assigned kuttl-test-tender-ghoul/production-ui-query-6699c7c8cb-2mj2k to ip-10-0-50-117.us-west-2.compute.internal default-scheduler logger.go:42: 07:44:50 | production | 2024-09-26 07:44:42 +0000 UTC Normal Pod production-ui-query-6699c7c8cb-2mj2k AddedInterface Add eth0 [10.130.0.80/23] from ovn-kubernetes logger.go:42: 07:44:50 | production | 2024-09-26 07:44:42 +0000 UTC Normal Pod production-ui-query-6699c7c8cb-2mj2k.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:44:42 +0000 UTC Normal Pod production-ui-query-6699c7c8cb-2mj2k.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:44:42 +0000 UTC Normal Pod production-ui-query-6699c7c8cb-2mj2k.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:44:42 +0000 UTC Normal Pod production-ui-query-6699c7c8cb-2mj2k.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:44:42 +0000 UTC Normal ReplicaSet.apps production-ui-query-6699c7c8cb SuccessfulCreate Created pod: production-ui-query-6699c7c8cb-2mj2k replicaset-controller logger.go:42: 07:44:50 | production | 2024-09-26 07:44:42 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-6699c7c8cb to 1 deployment-controller logger.go:42: 07:44:50 | production | 2024-09-26 07:44:43 +0000 UTC Normal Pod production-ui-query-6699c7c8cb-2mj2k.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:44:50 | production | 2024-09-26 07:44:43 +0000 UTC Normal Pod production-ui-query-6699c7c8cb-2mj2k.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:44:50 | production | Deleting namespace: kuttl-test-tender-ghoul === CONT kuttl/harness/artifacts logger.go:42: 07:44:56 | artifacts | Creating namespace: kuttl-test-stable-poodle logger.go:42: 07:44:56 | artifacts | artifacts events from ns kuttl-test-stable-poodle: logger.go:42: 07:44:56 | artifacts | Deleting namespace: kuttl-test-stable-poodle === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (148.74s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/allinone (40.76s) --- PASS: kuttl/harness/production (101.38s) --- PASS: kuttl/harness/artifacts (5.90s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name ui --report --output /logs/artifacts/ui.xml ./artifacts/kuttl-report.xml time="2024-09-26T07:45:02Z" level=debug msg="Setting a new name for the test suites" time="2024-09-26T07:45:02Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-09-26T07:45:02Z" level=debug msg="normalizing test case names" time="2024-09-26T07:45:02Z" level=debug msg="ui/allinone -> ui_allinone" time="2024-09-26T07:45:02Z" level=debug msg="ui/production -> ui_production" time="2024-09-26T07:45:02Z" level=debug msg="ui/artifacts -> ui_artifacts" +---------------+--------+ | NAME | RESULT | +---------------+--------+ | ui_allinone | passed | | ui_production | passed | | ui_artifacts | passed | +---------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests'